var/home/core/zuul-output/0000755000175000017500000000000015114544715014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114552261015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004737055515114552251017716 0ustar rootrootDec 05 12:06:09 crc systemd[1]: Starting Kubernetes Kubelet... Dec 05 12:06:09 crc restorecon[4690]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:09 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 12:06:10 crc restorecon[4690]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 05 12:06:11 crc kubenswrapper[4807]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 12:06:11 crc kubenswrapper[4807]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 12:06:11 crc kubenswrapper[4807]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 12:06:11 crc kubenswrapper[4807]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 12:06:11 crc kubenswrapper[4807]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 12:06:11 crc kubenswrapper[4807]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.056344 4807 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060245 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060272 4807 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060279 4807 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060287 4807 feature_gate.go:330] unrecognized feature gate: Example Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060293 4807 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060299 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060305 4807 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060311 4807 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060317 4807 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060323 4807 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060329 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060336 4807 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060343 4807 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060351 4807 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060360 4807 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060372 4807 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060377 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060383 4807 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060388 4807 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060394 4807 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060399 4807 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060404 4807 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060409 4807 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060415 4807 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060422 4807 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060428 4807 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060435 4807 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060441 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060448 4807 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060455 4807 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060461 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060466 4807 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060472 4807 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060477 4807 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060482 4807 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060487 4807 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060492 4807 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060499 4807 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060505 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060511 4807 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060517 4807 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060541 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060547 4807 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060553 4807 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060559 4807 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060565 4807 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060571 4807 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060578 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060584 4807 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060589 4807 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060594 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060600 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060606 4807 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060611 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060616 4807 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060622 4807 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060627 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060632 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060637 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060642 4807 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060647 4807 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060652 4807 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060657 4807 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060665 4807 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060672 4807 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060678 4807 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060683 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060690 4807 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060697 4807 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060702 4807 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.060708 4807 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061044 4807 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061090 4807 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061107 4807 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061117 4807 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061127 4807 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061137 4807 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061149 4807 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061159 4807 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061167 4807 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061175 4807 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061183 4807 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061190 4807 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061198 4807 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061206 4807 flags.go:64] FLAG: --cgroup-root="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061215 4807 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061223 4807 flags.go:64] FLAG: --client-ca-file="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061233 4807 flags.go:64] FLAG: --cloud-config="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061241 4807 flags.go:64] FLAG: --cloud-provider="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061249 4807 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061257 4807 flags.go:64] FLAG: --cluster-domain="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061264 4807 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061271 4807 flags.go:64] FLAG: --config-dir="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061277 4807 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061284 4807 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061292 4807 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061299 4807 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061306 4807 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061312 4807 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061318 4807 flags.go:64] FLAG: --contention-profiling="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061324 4807 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061331 4807 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061337 4807 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061343 4807 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061356 4807 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061363 4807 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061369 4807 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061376 4807 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061384 4807 flags.go:64] FLAG: --enable-server="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061391 4807 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061428 4807 flags.go:64] FLAG: --event-burst="100" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061438 4807 flags.go:64] FLAG: --event-qps="50" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061446 4807 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061455 4807 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061462 4807 flags.go:64] FLAG: --eviction-hard="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061473 4807 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061480 4807 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061487 4807 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061496 4807 flags.go:64] FLAG: --eviction-soft="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061504 4807 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061511 4807 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061519 4807 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061551 4807 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061560 4807 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061566 4807 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061572 4807 flags.go:64] FLAG: --feature-gates="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061581 4807 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061587 4807 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061595 4807 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061601 4807 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061608 4807 flags.go:64] FLAG: --healthz-port="10248" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061615 4807 flags.go:64] FLAG: --help="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061621 4807 flags.go:64] FLAG: --hostname-override="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061627 4807 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061634 4807 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061640 4807 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061646 4807 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061652 4807 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061658 4807 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061664 4807 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061670 4807 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061676 4807 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061685 4807 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061692 4807 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061698 4807 flags.go:64] FLAG: --kube-reserved="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061704 4807 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061710 4807 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061716 4807 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061722 4807 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061728 4807 flags.go:64] FLAG: --lock-file="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061734 4807 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061741 4807 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061747 4807 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061756 4807 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061762 4807 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061768 4807 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061774 4807 flags.go:64] FLAG: --logging-format="text" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061780 4807 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061787 4807 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061795 4807 flags.go:64] FLAG: --manifest-url="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061801 4807 flags.go:64] FLAG: --manifest-url-header="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061809 4807 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061816 4807 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061824 4807 flags.go:64] FLAG: --max-pods="110" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061830 4807 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061836 4807 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061842 4807 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061847 4807 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061854 4807 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061861 4807 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061867 4807 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061882 4807 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061888 4807 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061894 4807 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061901 4807 flags.go:64] FLAG: --pod-cidr="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061907 4807 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061917 4807 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061923 4807 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061929 4807 flags.go:64] FLAG: --pods-per-core="0" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061935 4807 flags.go:64] FLAG: --port="10250" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061943 4807 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061948 4807 flags.go:64] FLAG: --provider-id="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061954 4807 flags.go:64] FLAG: --qos-reserved="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061960 4807 flags.go:64] FLAG: --read-only-port="10255" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061966 4807 flags.go:64] FLAG: --register-node="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061972 4807 flags.go:64] FLAG: --register-schedulable="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061978 4807 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061989 4807 flags.go:64] FLAG: --registry-burst="10" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.061995 4807 flags.go:64] FLAG: --registry-qps="5" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062001 4807 flags.go:64] FLAG: --reserved-cpus="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062007 4807 flags.go:64] FLAG: --reserved-memory="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062015 4807 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062021 4807 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062027 4807 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062033 4807 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062040 4807 flags.go:64] FLAG: --runonce="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062047 4807 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062053 4807 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062059 4807 flags.go:64] FLAG: --seccomp-default="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062065 4807 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062071 4807 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062078 4807 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062085 4807 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062091 4807 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062097 4807 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062103 4807 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062110 4807 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062115 4807 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062122 4807 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062128 4807 flags.go:64] FLAG: --system-cgroups="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062134 4807 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062143 4807 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062149 4807 flags.go:64] FLAG: --tls-cert-file="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062155 4807 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062162 4807 flags.go:64] FLAG: --tls-min-version="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062167 4807 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062173 4807 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062180 4807 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062186 4807 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062191 4807 flags.go:64] FLAG: --v="2" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062200 4807 flags.go:64] FLAG: --version="false" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062208 4807 flags.go:64] FLAG: --vmodule="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062216 4807 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062223 4807 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062366 4807 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062373 4807 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062379 4807 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062384 4807 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062389 4807 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062395 4807 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062400 4807 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062406 4807 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062412 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062427 4807 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062434 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062441 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062448 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062454 4807 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062461 4807 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062467 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062472 4807 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062477 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062483 4807 feature_gate.go:330] unrecognized feature gate: Example Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062489 4807 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062494 4807 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062499 4807 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062504 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062509 4807 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062515 4807 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062520 4807 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062549 4807 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062554 4807 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062559 4807 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062565 4807 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062570 4807 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062576 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062583 4807 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062588 4807 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062594 4807 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062601 4807 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062609 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062615 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062621 4807 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062638 4807 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062645 4807 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062654 4807 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062660 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062668 4807 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062675 4807 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062681 4807 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062687 4807 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062692 4807 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062698 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062703 4807 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062710 4807 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062717 4807 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062723 4807 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062729 4807 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062735 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062741 4807 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062746 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062752 4807 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062759 4807 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062764 4807 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062770 4807 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062776 4807 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062782 4807 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062787 4807 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062793 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062799 4807 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062806 4807 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062812 4807 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062818 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062823 4807 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.062828 4807 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.062846 4807 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.071967 4807 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.072380 4807 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072473 4807 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072484 4807 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072490 4807 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072495 4807 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072500 4807 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072505 4807 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072510 4807 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072515 4807 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072541 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072546 4807 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072551 4807 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072555 4807 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072560 4807 feature_gate.go:330] unrecognized feature gate: Example Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072564 4807 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072570 4807 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072574 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072578 4807 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072583 4807 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072588 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072592 4807 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072597 4807 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072601 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072606 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072611 4807 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072617 4807 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072625 4807 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072631 4807 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072635 4807 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072640 4807 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072644 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072649 4807 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072655 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072662 4807 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072668 4807 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072673 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072678 4807 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072683 4807 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072687 4807 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072692 4807 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072696 4807 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072701 4807 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072706 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072710 4807 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072714 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072719 4807 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072723 4807 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072727 4807 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072732 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072736 4807 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072741 4807 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072747 4807 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072753 4807 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072758 4807 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072763 4807 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072769 4807 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072775 4807 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072781 4807 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072786 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072792 4807 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072796 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072801 4807 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072805 4807 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072810 4807 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072815 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072819 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072824 4807 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072828 4807 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072832 4807 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072836 4807 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072841 4807 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.072846 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.072856 4807 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073003 4807 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073013 4807 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073019 4807 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073025 4807 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073029 4807 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073034 4807 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073038 4807 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073044 4807 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073048 4807 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073052 4807 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073057 4807 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073062 4807 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073066 4807 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073099 4807 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073105 4807 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073111 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073115 4807 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073120 4807 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073127 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073131 4807 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073136 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073140 4807 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073145 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073150 4807 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073155 4807 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073160 4807 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073164 4807 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073169 4807 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073174 4807 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073178 4807 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073182 4807 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073187 4807 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073191 4807 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073198 4807 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073204 4807 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073208 4807 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073213 4807 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073219 4807 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073225 4807 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073231 4807 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073236 4807 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073241 4807 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073246 4807 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073251 4807 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073255 4807 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073260 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073264 4807 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073269 4807 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073273 4807 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073278 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073283 4807 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073288 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073292 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073297 4807 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073301 4807 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073307 4807 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073311 4807 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073316 4807 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073320 4807 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073325 4807 feature_gate.go:330] unrecognized feature gate: Example Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073329 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073333 4807 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073337 4807 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073342 4807 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073347 4807 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073351 4807 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073355 4807 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073360 4807 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073364 4807 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073368 4807 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.073373 4807 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.073381 4807 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.073853 4807 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.078172 4807 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.078263 4807 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.079305 4807 server.go:997] "Starting client certificate rotation" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.079321 4807 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.079627 4807 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-12 22:04:05.590606611 +0000 UTC Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.079735 4807 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 177h57m54.510876681s for next certificate rotation Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.086160 4807 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.087450 4807 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.104097 4807 log.go:25] "Validated CRI v1 runtime API" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.122230 4807 log.go:25] "Validated CRI v1 image API" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.124192 4807 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.126895 4807 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-05-12-01-32-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.126936 4807 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.155320 4807 manager.go:217] Machine: {Timestamp:2025-12-05 12:06:11.150522441 +0000 UTC m=+0.644385780 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:46e7cf4e-5f92-4473-b811-c45109d1ff45 BootID:b68900d3-0b24-4276-b2e2-04428fc986a2 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:d4:e5:e0 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:d4:e5:e0 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c3:5e:2c Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:27:e2:6e Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f9:22:41 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e7:93:1b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:0a:42:5d:c1:af:b6 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9a:7c:e5:8e:76:75 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.155738 4807 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.155914 4807 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.157057 4807 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.157467 4807 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.157574 4807 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.157959 4807 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.157977 4807 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.158252 4807 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.158297 4807 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.158880 4807 state_mem.go:36] "Initialized new in-memory state store" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.159052 4807 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.159885 4807 kubelet.go:418] "Attempting to sync node with API server" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.159930 4807 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.159961 4807 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.159987 4807 kubelet.go:324] "Adding apiserver pod source" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.160006 4807 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.162391 4807 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.165275 4807 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.169888 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.169993 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.169979 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.170101 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.170890 4807 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171779 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171831 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171849 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171869 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171900 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171930 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171949 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171971 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.171989 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.172002 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.172022 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.172036 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.172083 4807 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.172916 4807 server.go:1280] "Started kubelet" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.173190 4807 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.173242 4807 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.173868 4807 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.173877 4807 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.175665 4807 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.175719 4807 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 12:06:11 crc systemd[1]: Started Kubernetes Kubelet. Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.176025 4807 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 13:15:49.735778766 +0000 UTC Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.176751 4807 server.go:460] "Adding debug handlers to kubelet server" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.177997 4807 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.178057 4807 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.178079 4807 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.178182 4807 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.178788 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="200ms" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.178599 4807 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e50474eeca25f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 12:06:11.172844127 +0000 UTC m=+0.666707436,LastTimestamp:2025-12-05 12:06:11.172844127 +0000 UTC m=+0.666707436,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.178670 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.181050 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.181803 4807 factory.go:55] Registering systemd factory Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.181826 4807 factory.go:221] Registration of the systemd container factory successfully Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.182167 4807 factory.go:153] Registering CRI-O factory Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.182354 4807 factory.go:221] Registration of the crio container factory successfully Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.182659 4807 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.182848 4807 factory.go:103] Registering Raw factory Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.183011 4807 manager.go:1196] Started watching for new ooms in manager Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.185989 4807 manager.go:319] Starting recovery of all containers Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192339 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192504 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192571 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192603 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192629 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192656 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192680 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192707 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192740 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192765 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192790 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192820 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192844 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192877 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192903 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192927 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192957 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.192984 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193009 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193036 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193063 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193092 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193120 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193149 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193176 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193203 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193281 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193316 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193344 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193371 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193399 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193429 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193456 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193484 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193514 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193579 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193611 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193640 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193667 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193692 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193721 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193748 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193776 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193804 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193830 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193854 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193878 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193905 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193931 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193956 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.193984 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194013 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194049 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194082 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194111 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194140 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194171 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194197 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194223 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194312 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194342 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194369 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194398 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194425 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194451 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194477 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194502 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.194728 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197586 4807 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197704 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197745 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197773 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197807 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197834 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197861 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197914 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197947 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.197978 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198007 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198039 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198069 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198100 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198128 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198158 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198187 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198217 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198247 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198274 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198308 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198335 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198362 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198391 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198418 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198447 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198477 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198506 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198577 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198610 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198638 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198665 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198693 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198723 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198753 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198782 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198879 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198940 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.198979 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199013 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199045 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199076 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199107 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199136 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199168 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199200 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199234 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199267 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199295 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199325 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199351 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199378 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199411 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199438 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199469 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199498 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199566 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199600 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199630 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199659 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199689 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199718 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199746 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199776 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199806 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199834 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199862 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.199893 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200016 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200047 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200078 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200107 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200134 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200162 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200189 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200217 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200244 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200274 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200302 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200330 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200356 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200386 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200416 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200444 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200477 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200504 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200597 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200627 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200657 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200686 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200715 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200746 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200774 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200800 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200827 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200856 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200883 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200966 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.200997 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201028 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201058 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201086 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201114 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201141 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201168 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201199 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201227 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201286 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201314 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201340 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201375 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201407 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201436 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201465 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201492 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201519 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201590 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201619 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201646 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201672 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201699 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201773 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201806 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201833 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201858 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201884 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201911 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201939 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201966 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.201992 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202020 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202048 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202077 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202104 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202132 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202158 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202185 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202210 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202241 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202349 4807 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202376 4807 reconstruct.go:97] "Volume reconstruction finished" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.202394 4807 reconciler.go:26] "Reconciler: start to sync state" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.227826 4807 manager.go:324] Recovery completed Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.232317 4807 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.234072 4807 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.234117 4807 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.234165 4807 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.234224 4807 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 05 12:06:11 crc kubenswrapper[4807]: W1205 12:06:11.235137 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.235193 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.238604 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.240116 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.240184 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.240199 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.241084 4807 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.241108 4807 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.241133 4807 state_mem.go:36] "Initialized new in-memory state store" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.278565 4807 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.334588 4807 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.378863 4807 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.379696 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="400ms" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.479302 4807 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.535612 4807 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.580155 4807 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.681221 4807 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.752712 4807 policy_none.go:49] "None policy: Start" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.756750 4807 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.756786 4807 state_mem.go:35] "Initializing new in-memory state store" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.781054 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="800ms" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.781520 4807 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.813844 4807 manager.go:334] "Starting Device Plugin manager" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.814103 4807 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.814122 4807 server.go:79] "Starting device plugin registration server" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.814921 4807 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.814940 4807 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.815109 4807 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.815212 4807 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.815224 4807 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.823959 4807 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.916049 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.917363 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.917406 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.917417 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.917444 4807 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 12:06:11 crc kubenswrapper[4807]: E1205 12:06:11.918039 4807 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.936083 4807 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.936204 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.937264 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.937300 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.937312 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.937430 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.937611 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.937702 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.938403 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.938438 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.938449 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.938962 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.939454 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.939559 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.940988 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941014 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941030 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941140 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941198 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941225 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941140 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941343 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941362 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941427 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941498 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.941614 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.942665 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.942720 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.942738 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.943025 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.943870 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.943898 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.944028 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.944151 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.944194 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.944862 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.944892 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.944904 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.945070 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.945105 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.945120 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.945165 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.945204 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.946002 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.946038 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:11 crc kubenswrapper[4807]: I1205 12:06:11.946067 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.044886 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.044920 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.044940 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.044956 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.044972 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.044999 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045013 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045025 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045044 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045059 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045121 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045191 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045208 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045222 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.045247 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.118774 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.120515 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.120611 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.120624 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.120661 4807 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 12:06:12 crc kubenswrapper[4807]: E1205 12:06:12.121281 4807 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.146883 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.147107 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.147317 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.147201 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.147595 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.147714 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.147824 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148009 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148129 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148117 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148267 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148343 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148315 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148389 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148400 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148454 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148479 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148570 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148572 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148654 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148726 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148702 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148860 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148922 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148904 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148990 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.148996 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.149025 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.149106 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.149213 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.175463 4807 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.176389 4807 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 07:23:15.474331331 +0000 UTC Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.176496 4807 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 787h17m3.297839805s for next certificate rotation Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.218308 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:12 crc kubenswrapper[4807]: E1205 12:06:12.218446 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.280102 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.294138 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.299968 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:12 crc kubenswrapper[4807]: E1205 12:06:12.301383 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.308166 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-34386d2700f469f47fa0579259ca7c021da7150ce1c28a5adb80f70acdb33add WatchSource:0}: Error finding container 34386d2700f469f47fa0579259ca7c021da7150ce1c28a5adb80f70acdb33add: Status 404 returned error can't find the container with id 34386d2700f469f47fa0579259ca7c021da7150ce1c28a5adb80f70acdb33add Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.313299 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ebafaef6482afe1d9710cba356f8214897e03a164ee1bb3eb29cf5f6aab5ab3d WatchSource:0}: Error finding container ebafaef6482afe1d9710cba356f8214897e03a164ee1bb3eb29cf5f6aab5ab3d: Status 404 returned error can't find the container with id ebafaef6482afe1d9710cba356f8214897e03a164ee1bb3eb29cf5f6aab5ab3d Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.315818 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.335844 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.337697 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c424adcbaa922f0df7312f1d7d2ca3b7c2d2db14653ccaefa15048395a8d03e2 WatchSource:0}: Error finding container c424adcbaa922f0df7312f1d7d2ca3b7c2d2db14653ccaefa15048395a8d03e2: Status 404 returned error can't find the container with id c424adcbaa922f0df7312f1d7d2ca3b7c2d2db14653ccaefa15048395a8d03e2 Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.342934 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.352526 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-fa9c6f10df99d90edcc1c1a946c14ce6bbe2186769a3332a9719b3ac36b58c8d WatchSource:0}: Error finding container fa9c6f10df99d90edcc1c1a946c14ce6bbe2186769a3332a9719b3ac36b58c8d: Status 404 returned error can't find the container with id fa9c6f10df99d90edcc1c1a946c14ce6bbe2186769a3332a9719b3ac36b58c8d Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.365025 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b277b2918ab98c3ef41671af75d46dbd421dc8577584c2491d013faab293b67a WatchSource:0}: Error finding container b277b2918ab98c3ef41671af75d46dbd421dc8577584c2491d013faab293b67a: Status 404 returned error can't find the container with id b277b2918ab98c3ef41671af75d46dbd421dc8577584c2491d013faab293b67a Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.470836 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:12 crc kubenswrapper[4807]: E1205 12:06:12.470963 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.521779 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.522980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.523023 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.523035 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:12 crc kubenswrapper[4807]: I1205 12:06:12.523066 4807 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 12:06:12 crc kubenswrapper[4807]: E1205 12:06:12.523563 4807 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 05 12:06:12 crc kubenswrapper[4807]: E1205 12:06:12.582164 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="1.6s" Dec 05 12:06:12 crc kubenswrapper[4807]: W1205 12:06:12.738382 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:12 crc kubenswrapper[4807]: E1205 12:06:12.738489 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.175195 4807 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.243875 4807 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae" exitCode=0 Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.243952 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.244045 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ebafaef6482afe1d9710cba356f8214897e03a164ee1bb3eb29cf5f6aab5ab3d"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.244180 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.245217 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.245247 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.245259 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.246759 4807 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a" exitCode=0 Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.246862 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.246892 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"34386d2700f469f47fa0579259ca7c021da7150ce1c28a5adb80f70acdb33add"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.246986 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.247698 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.247784 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.247797 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.249233 4807 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f" exitCode=0 Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.249302 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.249336 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b277b2918ab98c3ef41671af75d46dbd421dc8577584c2491d013faab293b67a"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.249400 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.250077 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.250090 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.250098 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.252143 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.252167 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.252178 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.252188 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fa9c6f10df99d90edcc1c1a946c14ce6bbe2186769a3332a9719b3ac36b58c8d"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.254398 4807 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa" exitCode=0 Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.254425 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.254440 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c424adcbaa922f0df7312f1d7d2ca3b7c2d2db14653ccaefa15048395a8d03e2"} Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.254505 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.254988 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.255007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.255017 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.257682 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.258431 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.258472 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.258489 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.323800 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.325249 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.325292 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.325304 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:13 crc kubenswrapper[4807]: I1205 12:06:13.325332 4807 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 12:06:13 crc kubenswrapper[4807]: E1205 12:06:13.326059 4807 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.259575 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6726759440c5213ec731623e92b8fce4760be11e904289c1dae4a5103b5af62e"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.259704 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.260954 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.260987 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.260997 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.263432 4807 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad" exitCode=0 Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.263473 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.263571 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.264207 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.264228 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.264235 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.298891 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.298949 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.298959 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.299064 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.300634 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.300667 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.300675 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.303049 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.303099 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.303689 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.303714 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.303736 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.310716 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.310747 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.310761 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.310777 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.310789 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3"} Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.310887 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.311436 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.311461 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.311493 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.927128 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.928724 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.928785 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.928803 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:14 crc kubenswrapper[4807]: I1205 12:06:14.928840 4807 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.039992 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.315689 4807 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3" exitCode=0 Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.315983 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.316009 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.316048 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3"} Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.316277 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.316473 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.317387 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.317411 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.317422 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.317417 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.317459 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.317481 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.318580 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.318632 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.318649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:15 crc kubenswrapper[4807]: I1205 12:06:15.932649 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.322762 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a"} Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.322838 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a"} Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.322851 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.322856 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed"} Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323010 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0"} Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323037 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05"} Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323050 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323711 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323742 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323754 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323857 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323892 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:16 crc kubenswrapper[4807]: I1205 12:06:16.323909 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.058110 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.058380 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.060223 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.060272 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.060288 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.324991 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.325566 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.325756 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.325780 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.325788 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.326578 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.326605 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:17 crc kubenswrapper[4807]: I1205 12:06:17.326613 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:18 crc kubenswrapper[4807]: I1205 12:06:18.808907 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 05 12:06:18 crc kubenswrapper[4807]: I1205 12:06:18.809180 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:18 crc kubenswrapper[4807]: I1205 12:06:18.810731 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:18 crc kubenswrapper[4807]: I1205 12:06:18.810798 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:18 crc kubenswrapper[4807]: I1205 12:06:18.810813 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:20 crc kubenswrapper[4807]: I1205 12:06:20.255205 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 05 12:06:20 crc kubenswrapper[4807]: I1205 12:06:20.255430 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:20 crc kubenswrapper[4807]: I1205 12:06:20.256934 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:20 crc kubenswrapper[4807]: I1205 12:06:20.256979 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:20 crc kubenswrapper[4807]: I1205 12:06:20.256991 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:21 crc kubenswrapper[4807]: I1205 12:06:21.561832 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:21 crc kubenswrapper[4807]: I1205 12:06:21.562016 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:21 crc kubenswrapper[4807]: I1205 12:06:21.563054 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:21 crc kubenswrapper[4807]: I1205 12:06:21.563100 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:21 crc kubenswrapper[4807]: I1205 12:06:21.563113 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:21 crc kubenswrapper[4807]: E1205 12:06:21.824142 4807 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.009789 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.015766 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.112414 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.120112 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.337169 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.338286 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.338328 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:22 crc kubenswrapper[4807]: I1205 12:06:22.338339 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:23 crc kubenswrapper[4807]: I1205 12:06:23.341209 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:23 crc kubenswrapper[4807]: I1205 12:06:23.342271 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:23 crc kubenswrapper[4807]: I1205 12:06:23.342310 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:23 crc kubenswrapper[4807]: I1205 12:06:23.342320 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:23 crc kubenswrapper[4807]: I1205 12:06:23.465794 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.175484 4807 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 05 12:06:24 crc kubenswrapper[4807]: E1205 12:06:24.182875 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.261125 4807 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.261216 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.343259 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.344214 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.344252 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.344260 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:24 crc kubenswrapper[4807]: W1205 12:06:24.510340 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.510438 4807 trace.go:236] Trace[1666192055]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 12:06:14.508) (total time: 10001ms): Dec 05 12:06:24 crc kubenswrapper[4807]: Trace[1666192055]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:06:24.510) Dec 05 12:06:24 crc kubenswrapper[4807]: Trace[1666192055]: [10.00182422s] [10.00182422s] END Dec 05 12:06:24 crc kubenswrapper[4807]: E1205 12:06:24.510463 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 12:06:24 crc kubenswrapper[4807]: W1205 12:06:24.581462 4807 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.581606 4807 trace.go:236] Trace[1243458996]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 12:06:14.578) (total time: 10003ms): Dec 05 12:06:24 crc kubenswrapper[4807]: Trace[1243458996]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (12:06:24.581) Dec 05 12:06:24 crc kubenswrapper[4807]: Trace[1243458996]: [10.003077691s] [10.003077691s] END Dec 05 12:06:24 crc kubenswrapper[4807]: E1205 12:06:24.581636 4807 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.824661 4807 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.824747 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.836274 4807 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 12:06:24 crc kubenswrapper[4807]: I1205 12:06:24.836358 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 12:06:25 crc kubenswrapper[4807]: I1205 12:06:25.937162 4807 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]log ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]etcd ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/generic-apiserver-start-informers ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/priority-and-fairness-filter ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-apiextensions-informers ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-apiextensions-controllers ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/crd-informer-synced ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-system-namespaces-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 05 12:06:25 crc kubenswrapper[4807]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/bootstrap-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/start-kube-aggregator-informers ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/apiservice-registration-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/apiservice-discovery-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]autoregister-completion ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/apiservice-openapi-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 05 12:06:25 crc kubenswrapper[4807]: livez check failed Dec 05 12:06:25 crc kubenswrapper[4807]: I1205 12:06:25.937236 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:06:26 crc kubenswrapper[4807]: I1205 12:06:26.466361 4807 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 12:06:26 crc kubenswrapper[4807]: I1205 12:06:26.466458 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.058681 4807 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.168795 4807 apiserver.go:52] "Watching apiserver" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.172757 4807 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.173137 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.173544 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.173669 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.173687 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.173857 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.173947 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.174243 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.174363 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.174466 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.174517 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.175689 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.175835 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.175836 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.176031 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.176203 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.176242 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.176431 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.176596 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.177184 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.179127 4807 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.202408 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.217897 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.232992 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.246304 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.257198 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.266846 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.279870 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.823458 4807 trace.go:236] Trace[1618442639]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 12:06:15.340) (total time: 14482ms): Dec 05 12:06:29 crc kubenswrapper[4807]: Trace[1618442639]: ---"Objects listed" error: 14482ms (12:06:29.823) Dec 05 12:06:29 crc kubenswrapper[4807]: Trace[1618442639]: [14.482735167s] [14.482735167s] END Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.823498 4807 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.823780 4807 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.823904 4807 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.830491 4807 trace.go:236] Trace[1093377998]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 12:06:14.872) (total time: 14958ms): Dec 05 12:06:29 crc kubenswrapper[4807]: Trace[1093377998]: ---"Objects listed" error: 14958ms (12:06:29.830) Dec 05 12:06:29 crc kubenswrapper[4807]: Trace[1093377998]: [14.958141308s] [14.958141308s] END Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.830518 4807 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.858334 4807 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.924939 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925216 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925323 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925425 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925493 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925605 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925241 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925689 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925469 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925683 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925769 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925808 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925908 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.925966 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926056 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926093 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926122 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926142 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926148 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926158 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926177 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926196 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926212 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926231 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926247 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926246 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926264 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926280 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926296 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926313 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926329 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926346 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926361 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926379 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926397 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926424 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926425 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926441 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926456 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926473 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926489 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926504 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926535 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926552 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926567 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926584 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926599 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926615 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926615 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926630 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926647 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926664 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926682 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926699 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926718 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926733 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926748 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926755 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926772 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926790 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926817 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926833 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926847 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926863 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926877 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926892 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926908 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926923 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926942 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926958 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926977 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926999 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927019 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927039 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927057 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927072 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927088 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927105 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927121 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927138 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927156 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927201 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927224 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927241 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927257 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927273 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927291 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927308 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927325 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927341 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927359 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927374 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927390 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927407 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927424 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927440 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927456 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927472 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927490 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929692 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929720 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929739 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929761 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929779 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929796 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929813 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929830 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929868 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929885 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929903 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930101 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930134 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930152 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930169 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930188 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930205 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930231 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930253 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930277 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930294 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930311 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930327 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930343 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930359 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930978 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931008 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931027 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931044 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931062 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931081 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931098 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931116 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931134 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931154 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931175 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931192 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931210 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931228 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931244 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931263 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931279 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931297 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931315 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931333 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931350 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931368 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931386 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931404 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931422 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931440 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931459 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931475 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931493 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931510 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931542 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931559 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931576 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931593 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931615 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931633 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931652 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931669 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931688 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931707 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931726 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931744 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931762 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931780 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931798 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931815 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931834 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931852 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931869 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931887 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931905 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931923 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931941 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931962 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931979 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931997 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932015 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932033 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932054 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932073 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932093 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932113 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932132 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932149 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932218 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932238 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932257 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932275 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932294 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932313 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932386 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932406 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932424 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932442 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932461 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932480 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932499 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932518 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932553 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932572 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932590 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932609 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932652 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932677 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932697 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932718 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932738 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932758 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932776 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932796 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932815 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932835 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932862 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932884 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932904 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.932940 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933028 4807 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933042 4807 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933056 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933068 4807 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933082 4807 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933096 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933110 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933121 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933133 4807 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933145 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.933157 4807 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.926891 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927012 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927134 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927136 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927270 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927280 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927400 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927546 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927582 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927749 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927775 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927801 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927823 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927931 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927956 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927973 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.927978 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928073 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928106 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928119 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928099 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928145 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928309 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928309 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928315 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928331 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928395 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928422 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928465 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928492 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928540 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928557 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928576 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928675 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928694 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928707 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928732 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928747 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928855 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928919 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.928994 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929013 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929017 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929056 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929099 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929140 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929152 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929192 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.929199 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930080 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930154 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930191 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930256 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930514 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930567 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930577 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.930739 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.931334 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.933248 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.934356 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:30.434061702 +0000 UTC m=+19.927924971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.936867 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.936957 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.937086 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.937343 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.937452 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.937478 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.937924 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.938168 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.938209 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.938349 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.938711 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.938828 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.939549 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.939691 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.939744 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.939886 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.940031 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.940245 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.940305 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.940601 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.940634 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.940780 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.940878 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.941172 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.941298 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.941419 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.941661 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.941787 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.942049 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.942121 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.942272 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.942438 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.942814 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.942982 4807 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50832->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.943034 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.943040 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50832->192.168.126.11:17697: read: connection reset by peer" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.943202 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.943271 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.943377 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.943894 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.944752 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.945563 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.946492 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.947032 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.947047 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.947732 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.948225 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.948587 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.948739 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.948755 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.949151 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.949337 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.949596 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.949727 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.949766 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.949996 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950042 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950061 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950105 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950153 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950548 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950559 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950571 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950832 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950945 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.950967 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.951015 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.951032 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.951172 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.951294 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.951663 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:06:30.451635484 +0000 UTC m=+19.945498943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.951716 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.951762 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.952032 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.952032 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.952302 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.952673 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.952852 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:30.452813571 +0000 UTC m=+19.946677030 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.953355 4807 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.953984 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.954010 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.954083 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.954235 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.955013 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.955170 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.956238 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.960969 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.961746 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.961639 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.967092 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.967501 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.968344 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.969667 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.970373 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.970471 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.970572 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.970703 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:30.470679978 +0000 UTC m=+19.964543457 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.971116 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.971211 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.971280 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:29 crc kubenswrapper[4807]: E1205 12:06:29.971430 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:30.471417565 +0000 UTC m=+19.965281054 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.972575 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.973064 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.977089 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.977373 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.980626 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.981069 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.981149 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.983327 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.983331 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.983417 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.983679 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.984057 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.984138 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.984510 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.984538 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.984797 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.985414 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.985375 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.985626 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986037 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986117 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.985489 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986051 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986147 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986822 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986367 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.985909 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986518 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986583 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986866 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.986623 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.987274 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.987512 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.987575 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.987595 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.987739 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.987864 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.987922 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.989952 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.990098 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.992216 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.992386 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4807]: I1205 12:06:29.992978 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.000406 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.005428 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-bnj26"] Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.006420 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.008829 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.008970 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.008992 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.013901 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.017275 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.019303 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034821 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034867 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034944 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034960 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034971 4807 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034982 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034992 4807 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035003 4807 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035014 4807 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.034765 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035096 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035030 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035213 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035330 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035356 4807 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035370 4807 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035384 4807 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035395 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035406 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035416 4807 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035427 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035440 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035458 4807 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035473 4807 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035485 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035497 4807 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035512 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035540 4807 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035551 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035561 4807 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035572 4807 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035584 4807 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035594 4807 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035609 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035619 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035629 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035644 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035656 4807 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035666 4807 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035676 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035685 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035695 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035705 4807 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035714 4807 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035724 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035733 4807 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035743 4807 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035753 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035763 4807 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035780 4807 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035793 4807 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035826 4807 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035837 4807 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035849 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035861 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035874 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035887 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035902 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035915 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035925 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035936 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035946 4807 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035956 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035969 4807 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035979 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.035989 4807 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036004 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036013 4807 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036024 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036034 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036044 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036053 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036064 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036073 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036084 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036094 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036106 4807 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036118 4807 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036129 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036140 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036154 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036167 4807 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036183 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036197 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036213 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036223 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036236 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036248 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036264 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036278 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036293 4807 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036309 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036323 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036336 4807 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036351 4807 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036362 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036375 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036385 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036397 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036408 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036419 4807 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036429 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036439 4807 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036450 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036461 4807 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036472 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036482 4807 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036493 4807 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036503 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036515 4807 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036546 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036557 4807 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036567 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036578 4807 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036588 4807 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036598 4807 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036607 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036617 4807 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036628 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036638 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036648 4807 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036658 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036667 4807 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036680 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036692 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036703 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036715 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036725 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036737 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036747 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036757 4807 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036768 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036778 4807 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036788 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036798 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036809 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036818 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036829 4807 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036838 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036848 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036859 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036869 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036878 4807 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036889 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036898 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036909 4807 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036919 4807 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036929 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036938 4807 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036949 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036959 4807 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036969 4807 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036979 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036989 4807 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.036998 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037009 4807 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037024 4807 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037035 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037048 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037060 4807 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037071 4807 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037080 4807 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037091 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037104 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037117 4807 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037126 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037137 4807 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037148 4807 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037157 4807 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037166 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037178 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037188 4807 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037199 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037208 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037218 4807 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037230 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037239 4807 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037250 4807 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037260 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037269 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037278 4807 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037288 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037297 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037306 4807 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037315 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037323 4807 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037332 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.037341 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.045831 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.064824 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.076935 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.091131 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.094299 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.100077 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.102873 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.109558 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.139944 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9e5318b9-e4e5-42ff-acef-811e7440033b-hosts-file\") pod \"node-resolver-bnj26\" (UID: \"9e5318b9-e4e5-42ff-acef-811e7440033b\") " pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.139980 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpnfz\" (UniqueName: \"kubernetes.io/projected/9e5318b9-e4e5-42ff-acef-811e7440033b-kube-api-access-xpnfz\") pod \"node-resolver-bnj26\" (UID: \"9e5318b9-e4e5-42ff-acef-811e7440033b\") " pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.240839 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9e5318b9-e4e5-42ff-acef-811e7440033b-hosts-file\") pod \"node-resolver-bnj26\" (UID: \"9e5318b9-e4e5-42ff-acef-811e7440033b\") " pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.240891 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpnfz\" (UniqueName: \"kubernetes.io/projected/9e5318b9-e4e5-42ff-acef-811e7440033b-kube-api-access-xpnfz\") pod \"node-resolver-bnj26\" (UID: \"9e5318b9-e4e5-42ff-acef-811e7440033b\") " pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.241100 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9e5318b9-e4e5-42ff-acef-811e7440033b-hosts-file\") pod \"node-resolver-bnj26\" (UID: \"9e5318b9-e4e5-42ff-acef-811e7440033b\") " pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.258470 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpnfz\" (UniqueName: \"kubernetes.io/projected/9e5318b9-e4e5-42ff-acef-811e7440033b-kube-api-access-xpnfz\") pod \"node-resolver-bnj26\" (UID: \"9e5318b9-e4e5-42ff-acef-811e7440033b\") " pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.276181 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.290993 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.295062 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.295495 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.304855 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.316951 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.319005 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bnj26" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.331335 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: W1205 12:06:30.333155 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e5318b9_e4e5_42ff_acef_811e7440033b.slice/crio-4cb2d3b92581637c61607f621a3efdea439d06d7963272e2e23bbdf3abfcd9b7 WatchSource:0}: Error finding container 4cb2d3b92581637c61607f621a3efdea439d06d7963272e2e23bbdf3abfcd9b7: Status 404 returned error can't find the container with id 4cb2d3b92581637c61607f621a3efdea439d06d7963272e2e23bbdf3abfcd9b7 Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.343692 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.355565 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.359444 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950"} Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.359541 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd"} Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.359574 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"184ba1749980ffbfe3279a729b3f1ccbb580cf35e85c08e3ee858963d60d6456"} Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.365132 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135"} Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.365198 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"81edc70049b1540f7c2e5dd23cc5c83799857155910c0911495314f62e738490"} Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.367329 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.368538 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.369812 4807 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0" exitCode=255 Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.369894 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0"} Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.375167 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bnj26" event={"ID":"9e5318b9-e4e5-42ff-acef-811e7440033b","Type":"ContainerStarted","Data":"4cb2d3b92581637c61607f621a3efdea439d06d7963272e2e23bbdf3abfcd9b7"} Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.376950 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2681ff3ec56c3dba8c976b6d7e2dc8c7f7600ded6811c4dd11a269df68b8ac3c"} Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.385687 4807 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.401138 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.401345 4807 scope.go:117] "RemoveContainer" containerID="054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.401481 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.419804 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.442779 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.443761 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.443848 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:31.443819313 +0000 UTC m=+20.937682582 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.453860 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.474155 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.491511 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.504278 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.515544 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.524268 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.541662 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.543811 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.543864 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.543892 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544078 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:06:31.544040443 +0000 UTC m=+21.037903712 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544138 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544262 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:31.544229717 +0000 UTC m=+21.038093036 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544262 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544279 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544286 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544298 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544304 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544313 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544355 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:31.544339609 +0000 UTC m=+21.038202878 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:30 crc kubenswrapper[4807]: E1205 12:06:30.544371 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:31.54436513 +0000 UTC m=+21.038228399 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.544392 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.553036 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.568982 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.577942 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.603911 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.621936 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:30Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.633749 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:30Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.648485 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:30Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.661077 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:30Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.936659 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-wd694"] Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.937022 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.939666 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.941586 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.941715 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.942841 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.943014 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.967437 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:30Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:30 crc kubenswrapper[4807]: I1205 12:06:30.990546 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:30Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.020722 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.042268 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.047939 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl8hc\" (UniqueName: \"kubernetes.io/projected/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-kube-api-access-hl8hc\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.047992 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-host\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.048008 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-serviceca\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.063643 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.074807 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.085201 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.095814 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.106089 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.117251 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.129471 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.144683 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.148947 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl8hc\" (UniqueName: \"kubernetes.io/projected/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-kube-api-access-hl8hc\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.148992 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-host\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.149008 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-serviceca\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.149219 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-host\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.149840 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-serviceca\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.166734 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.167003 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl8hc\" (UniqueName: \"kubernetes.io/projected/c1360d4c-bc7f-4557-9267-5fcfc052d9ea-kube-api-access-hl8hc\") pod \"node-ca-wd694\" (UID: \"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\") " pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.178179 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.190215 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.201788 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.219561 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.235012 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.235038 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.235189 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.235352 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.235611 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.235697 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.239028 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.239863 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.240749 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.241710 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.242418 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.243162 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.243706 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.243966 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.244663 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.245475 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.248211 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.249054 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.249335 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wd694" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.256166 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.257020 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.260583 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.260815 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.263886 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.264413 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.265609 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.266039 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.267076 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.267763 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.268240 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.269183 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.269636 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.270662 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.271103 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.272130 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.272949 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.273416 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: W1205 12:06:31.272961 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1360d4c_bc7f_4557_9267_5fcfc052d9ea.slice/crio-79e549209fe51d7042bfb936273972f2d0781ee61506c6c92cc266c52315d20d WatchSource:0}: Error finding container 79e549209fe51d7042bfb936273972f2d0781ee61506c6c92cc266c52315d20d: Status 404 returned error can't find the container with id 79e549209fe51d7042bfb936273972f2d0781ee61506c6c92cc266c52315d20d Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.274329 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.274826 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.275697 4807 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.275792 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.277738 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.280245 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.280817 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.281107 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.283397 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.284250 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.285193 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.285829 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.286919 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.287501 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.288515 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.289608 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.290181 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.290998 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.291512 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.292920 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.293744 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.294463 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.295343 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.295825 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.296817 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.297443 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.298300 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.319214 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.335294 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.357573 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.370286 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.379998 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wd694" event={"ID":"c1360d4c-bc7f-4557-9267-5fcfc052d9ea","Type":"ContainerStarted","Data":"79e549209fe51d7042bfb936273972f2d0781ee61506c6c92cc266c52315d20d"} Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.382139 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.383107 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.385600 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928"} Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.386192 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.391725 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bnj26" event={"ID":"9e5318b9-e4e5-42ff-acef-811e7440033b","Type":"ContainerStarted","Data":"7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268"} Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.393313 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.398485 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.410931 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.430205 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.447556 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.450668 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.451158 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.451206 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:33.4511937 +0000 UTC m=+22.945056969 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.462425 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.476777 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.495304 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.509868 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.529921 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.551928 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.552051 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552145 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:06:33.552108054 +0000 UTC m=+23.045971333 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552197 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552223 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552239 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.552249 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.552288 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552335 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:33.552324679 +0000 UTC m=+23.046188058 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552421 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552478 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:33.552467422 +0000 UTC m=+23.046330691 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552479 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552517 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552544 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:31 crc kubenswrapper[4807]: E1205 12:06:31.552614 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:33.552591695 +0000 UTC m=+23.046455024 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.554460 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.566385 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.580072 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.597703 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.608638 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.621074 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.752823 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-tsqzx"] Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.753460 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.760351 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.760767 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.760945 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.761439 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.765306 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-kth9r"] Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.765740 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.766687 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.771334 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.776915 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.776916 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.777363 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.778728 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-sq7cs"] Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.779147 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.787612 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.788985 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.789067 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.795545 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.809068 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.837236 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.850925 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854297 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-socket-dir-parent\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854329 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-hostroot\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854343 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-conf-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854357 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-multus-certs\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854379 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-system-cni-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854404 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cnibin\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854421 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854435 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-k8s-cni-cncf-io\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854449 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a779882f-2b7e-4ae0-addd-686fd4343bb6-rootfs\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854469 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j59f\" (UniqueName: \"kubernetes.io/projected/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-kube-api-access-5j59f\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854484 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-cni-bin\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854506 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a779882f-2b7e-4ae0-addd-686fd4343bb6-mcd-auth-proxy-config\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854544 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cni-binary-copy\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854562 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854585 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-os-release\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854601 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-kubelet\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854617 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-netns\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854632 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/379e7e5d-522d-409b-84db-fafbd1bff182-cni-binary-copy\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854646 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-cni-multus\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854662 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/379e7e5d-522d-409b-84db-fafbd1bff182-multus-daemon-config\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854679 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwdrs\" (UniqueName: \"kubernetes.io/projected/a779882f-2b7e-4ae0-addd-686fd4343bb6-kube-api-access-pwdrs\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854694 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-cni-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854707 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-cnibin\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854721 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-os-release\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854737 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a779882f-2b7e-4ae0-addd-686fd4343bb6-proxy-tls\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854753 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-system-cni-dir\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854768 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-etc-kubernetes\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.854782 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svt9b\" (UniqueName: \"kubernetes.io/projected/379e7e5d-522d-409b-84db-fafbd1bff182-kube-api-access-svt9b\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.885051 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.907626 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.925640 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.937629 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.949191 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955677 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-socket-dir-parent\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955724 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-system-cni-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955753 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-hostroot\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955777 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-conf-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955794 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-multus-certs\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955826 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cnibin\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955847 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955870 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-k8s-cni-cncf-io\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955860 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-socket-dir-parent\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955898 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cnibin\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955911 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a779882f-2b7e-4ae0-addd-686fd4343bb6-rootfs\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955872 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-multus-certs\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955908 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-conf-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955947 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j59f\" (UniqueName: \"kubernetes.io/projected/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-kube-api-access-5j59f\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955955 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a779882f-2b7e-4ae0-addd-686fd4343bb6-rootfs\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955968 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-cni-bin\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955864 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-system-cni-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955986 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a779882f-2b7e-4ae0-addd-686fd4343bb6-mcd-auth-proxy-config\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.955937 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-k8s-cni-cncf-io\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956003 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cni-binary-copy\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956011 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-cni-bin\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956033 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956060 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-os-release\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956004 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-hostroot\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956091 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-kubelet\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956152 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-netns\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956178 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwdrs\" (UniqueName: \"kubernetes.io/projected/a779882f-2b7e-4ae0-addd-686fd4343bb6-kube-api-access-pwdrs\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956193 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-cni-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956206 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-cnibin\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956220 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/379e7e5d-522d-409b-84db-fafbd1bff182-cni-binary-copy\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956234 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-cni-multus\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956249 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/379e7e5d-522d-409b-84db-fafbd1bff182-multus-daemon-config\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956265 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-os-release\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956281 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a779882f-2b7e-4ae0-addd-686fd4343bb6-proxy-tls\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956295 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-etc-kubernetes\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956310 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svt9b\" (UniqueName: \"kubernetes.io/projected/379e7e5d-522d-409b-84db-fafbd1bff182-kube-api-access-svt9b\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956324 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-system-cni-dir\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956384 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-system-cni-dir\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956422 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956554 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-kubelet\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956562 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-multus-cni-dir\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956582 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-run-netns\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956605 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-os-release\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956654 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a779882f-2b7e-4ae0-addd-686fd4343bb6-mcd-auth-proxy-config\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956691 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cni-binary-copy\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956700 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-cnibin\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.956717 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-etc-kubernetes\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.957102 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/379e7e5d-522d-409b-84db-fafbd1bff182-cni-binary-copy\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.957125 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/379e7e5d-522d-409b-84db-fafbd1bff182-multus-daemon-config\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.957136 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/379e7e5d-522d-409b-84db-fafbd1bff182-host-var-lib-cni-multus\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.957176 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-os-release\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.957576 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.962609 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a779882f-2b7e-4ae0-addd-686fd4343bb6-proxy-tls\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.967158 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.975298 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwdrs\" (UniqueName: \"kubernetes.io/projected/a779882f-2b7e-4ae0-addd-686fd4343bb6-kube-api-access-pwdrs\") pod \"machine-config-daemon-kth9r\" (UID: \"a779882f-2b7e-4ae0-addd-686fd4343bb6\") " pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.975354 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j59f\" (UniqueName: \"kubernetes.io/projected/f23809b6-3d63-484f-8e7c-9dc5dac87ebd-kube-api-access-5j59f\") pod \"multus-additional-cni-plugins-tsqzx\" (UID: \"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\") " pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.979060 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svt9b\" (UniqueName: \"kubernetes.io/projected/379e7e5d-522d-409b-84db-fafbd1bff182-kube-api-access-svt9b\") pod \"multus-sq7cs\" (UID: \"379e7e5d-522d-409b-84db-fafbd1bff182\") " pod="openshift-multus/multus-sq7cs" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.982423 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:31 crc kubenswrapper[4807]: I1205 12:06:31.996503 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.008999 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.042328 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.066440 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.081001 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:06:32 crc kubenswrapper[4807]: W1205 12:06:32.081957 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf23809b6_3d63_484f_8e7c_9dc5dac87ebd.slice/crio-fbc82a0feb1d9bac57ae5f260161775149bcde317b90bda735adc2d5afd528c5 WatchSource:0}: Error finding container fbc82a0feb1d9bac57ae5f260161775149bcde317b90bda735adc2d5afd528c5: Status 404 returned error can't find the container with id fbc82a0feb1d9bac57ae5f260161775149bcde317b90bda735adc2d5afd528c5 Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.083255 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.091414 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sq7cs" Dec 05 12:06:32 crc kubenswrapper[4807]: W1205 12:06:32.091946 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda779882f_2b7e_4ae0_addd_686fd4343bb6.slice/crio-ca42fb72331f20cc3eaec97f972a9e33d410ef776037f0aa0413dfa12beaeee4 WatchSource:0}: Error finding container ca42fb72331f20cc3eaec97f972a9e33d410ef776037f0aa0413dfa12beaeee4: Status 404 returned error can't find the container with id ca42fb72331f20cc3eaec97f972a9e33d410ef776037f0aa0413dfa12beaeee4 Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.124261 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.144149 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-84bd8"] Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.150753 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.160397 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.172567 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.192575 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.212887 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.231775 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.252730 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.258856 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-slash\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.258907 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-script-lib\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.258963 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-systemd\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.258982 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g748d\" (UniqueName: \"kubernetes.io/projected/a9bb753e-da01-4d90-95b2-5614ab510ec8-kube-api-access-g748d\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259073 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-netd\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259121 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-log-socket\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259159 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-systemd-units\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259175 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-node-log\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259194 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-env-overrides\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259215 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-etc-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259249 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259272 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-ovn-kubernetes\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259295 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-netns\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259313 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-bin\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259329 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovn-node-metrics-cert\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259346 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259367 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-kubelet\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259384 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-var-lib-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259399 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-ovn\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.259412 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-config\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.273208 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.292995 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.341226 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360444 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360509 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-ovn-kubernetes\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360560 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovn-node-metrics-cert\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360587 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-netns\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360606 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-bin\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360625 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360666 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360666 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-bin\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360640 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-ovn-kubernetes\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360637 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360682 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-netns\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360929 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-kubelet\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.360999 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-config\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361033 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-var-lib-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361059 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-ovn\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361090 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-slash\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361086 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-kubelet\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361115 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-script-lib\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361171 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-slash\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361198 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-var-lib-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361223 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-systemd\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361259 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-ovn\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361267 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g748d\" (UniqueName: \"kubernetes.io/projected/a9bb753e-da01-4d90-95b2-5614ab510ec8-kube-api-access-g748d\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361323 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-netd\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361343 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-systemd\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361357 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-log-socket\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361376 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-netd\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361412 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-systemd-units\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361454 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-node-log\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361474 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-systemd-units\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361482 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-env-overrides\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361511 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-etc-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361518 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-node-log\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361441 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-log-socket\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.361598 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-etc-openvswitch\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.362041 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-config\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.362071 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-script-lib\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.362080 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-env-overrides\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.367831 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovn-node-metrics-cert\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.386005 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.396550 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.398706 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wd694" event={"ID":"c1360d4c-bc7f-4557-9267-5fcfc052d9ea","Type":"ContainerStarted","Data":"74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.400767 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.400812 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.400834 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"ca42fb72331f20cc3eaec97f972a9e33d410ef776037f0aa0413dfa12beaeee4"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.402171 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerStarted","Data":"dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.402218 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerStarted","Data":"fbc82a0feb1d9bac57ae5f260161775149bcde317b90bda735adc2d5afd528c5"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.404389 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerStarted","Data":"3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.404466 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerStarted","Data":"ec43ad49d2c75b89ef93518e4a0767e9e187b5cbc50ecb3180ac610e9b9ad783"} Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.412970 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g748d\" (UniqueName: \"kubernetes.io/projected/a9bb753e-da01-4d90-95b2-5614ab510ec8-kube-api-access-g748d\") pod \"ovnkube-node-84bd8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.438767 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.465850 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:32 crc kubenswrapper[4807]: W1205 12:06:32.484266 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9bb753e_da01_4d90_95b2_5614ab510ec8.slice/crio-72e81ff6b8914de874641ae53375e42e97c220d16b2d2d1bef37443420f0d19e WatchSource:0}: Error finding container 72e81ff6b8914de874641ae53375e42e97c220d16b2d2d1bef37443420f0d19e: Status 404 returned error can't find the container with id 72e81ff6b8914de874641ae53375e42e97c220d16b2d2d1bef37443420f0d19e Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.486101 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.524382 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.573125 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.602341 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.642509 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.680486 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.722017 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.761067 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.800107 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.843081 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.882932 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.927367 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:32 crc kubenswrapper[4807]: I1205 12:06:32.962780 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.023859 4807 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.026309 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.026352 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.026423 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.026901 4807 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.035861 4807 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.036123 4807 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.037355 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.037398 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.037413 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.037432 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.037443 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.055705 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.073094 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.077094 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.077131 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.077142 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.077157 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.077165 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.081836 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.090906 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.095238 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.095274 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.095288 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.095306 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.095317 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.107730 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.110816 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.110848 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.110859 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.110874 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.110885 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.119688 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.121760 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.124839 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.124875 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.124884 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.124900 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.124909 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.135762 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.135873 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.137294 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.137327 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.137338 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.137354 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.137363 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.160879 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.202975 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.235351 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.235358 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.235512 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.235629 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.235767 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.236059 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.239189 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.239229 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.239238 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.239252 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.239262 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.340986 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.341020 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.341033 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.341049 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.341058 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.407729 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3" exitCode=0 Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.408040 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.408230 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"72e81ff6b8914de874641ae53375e42e97c220d16b2d2d1bef37443420f0d19e"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.409388 4807 generic.go:334] "Generic (PLEG): container finished" podID="f23809b6-3d63-484f-8e7c-9dc5dac87ebd" containerID="dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376" exitCode=0 Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.409408 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerDied","Data":"dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.438410 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.443741 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.443776 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.443785 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.443800 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.443831 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.452900 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.468122 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.475799 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.480559 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.481478 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:37.481451976 +0000 UTC m=+26.975315255 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.481669 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.485641 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.487949 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.489285 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.502921 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.514479 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.534168 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.548210 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.548269 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.548288 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.548315 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.548332 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.566483 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.582580 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.582823 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.582945 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.583034 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:06:37.583002504 +0000 UTC m=+27.076865933 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.583071 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.583095 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.583107 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.583173 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:37.583151409 +0000 UTC m=+27.077014678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.583210 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.583367 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.583427 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.583457 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:37.583448585 +0000 UTC m=+27.077311854 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.584352 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.584382 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.584394 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:33 crc kubenswrapper[4807]: E1205 12:06:33.584437 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:37.584425287 +0000 UTC m=+27.078288736 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.630412 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.650971 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.651005 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.651014 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.651028 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.651038 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.662502 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.702094 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.744220 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.753372 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.753415 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.753424 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.753440 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.753449 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.794566 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.822269 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.855723 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.855771 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.855783 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.855802 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.855819 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.863167 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.899337 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.940362 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.960782 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.960837 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.960876 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.960904 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.960918 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:33Z","lastTransitionTime":"2025-12-05T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:33 crc kubenswrapper[4807]: I1205 12:06:33.986202 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:33Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.020213 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.064063 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.064102 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.064113 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.064129 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.064140 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.068632 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.104099 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.154320 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.166508 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.166557 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.166568 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.166586 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.166596 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.186408 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.221709 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.262031 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.269104 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.269145 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.269153 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.269179 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.269192 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.303088 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.341890 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.372164 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.372205 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.372217 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.372233 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.372246 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.382397 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.414537 4807 generic.go:334] "Generic (PLEG): container finished" podID="f23809b6-3d63-484f-8e7c-9dc5dac87ebd" containerID="91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5" exitCode=0 Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.414652 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerDied","Data":"91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.418575 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.418630 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.418649 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.418661 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.418674 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.418685 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.433177 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.463289 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.474315 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.474359 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.474370 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.474389 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.474400 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.502203 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.540321 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.577384 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.577763 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.577776 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.577800 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.577812 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.581405 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.621040 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.661240 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.680032 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.680082 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.680095 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.680113 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.680124 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.701767 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.740004 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.782600 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.782646 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.782662 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.782685 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.782709 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.786661 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.831713 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.862700 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.885007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.885052 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.885065 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.885083 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.885095 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.900704 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.945154 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.982320 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.987131 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.987168 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.987185 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.987206 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:34 crc kubenswrapper[4807]: I1205 12:06:34.987217 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:34Z","lastTransitionTime":"2025-12-05T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.089439 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.089494 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.089512 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.089568 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.089588 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.192504 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.192583 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.192594 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.192611 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.192623 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.234488 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.234593 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:35 crc kubenswrapper[4807]: E1205 12:06:35.234675 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.234776 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:35 crc kubenswrapper[4807]: E1205 12:06:35.234936 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:35 crc kubenswrapper[4807]: E1205 12:06:35.235050 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.295104 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.295149 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.295162 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.295177 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.295187 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.397847 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.397895 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.397940 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.397958 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.397969 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.426302 4807 generic.go:334] "Generic (PLEG): container finished" podID="f23809b6-3d63-484f-8e7c-9dc5dac87ebd" containerID="6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d" exitCode=0 Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.426348 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerDied","Data":"6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.444094 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.459206 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.473176 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.487487 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.500725 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.500766 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.500806 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.500824 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.500835 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.501941 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.511693 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.524738 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.536073 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.556313 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.568063 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.585752 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.598400 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.603136 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.603162 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.603172 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.603186 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.603197 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.611729 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.623745 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.636736 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:35Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.705590 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.705630 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.705643 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.705660 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.705672 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.807820 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.807865 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.807875 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.807894 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.807905 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.910439 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.910485 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.910493 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.910508 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:35 crc kubenswrapper[4807]: I1205 12:06:35.910518 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:35Z","lastTransitionTime":"2025-12-05T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.014088 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.014131 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.014139 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.014154 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.014164 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.116109 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.116152 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.116162 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.116180 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.116192 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.219251 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.219290 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.219300 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.219316 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.219327 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.321990 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.322040 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.322051 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.322069 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.322082 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.424574 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.424632 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.424645 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.424668 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.424682 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.431245 4807 generic.go:334] "Generic (PLEG): container finished" podID="f23809b6-3d63-484f-8e7c-9dc5dac87ebd" containerID="3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c" exitCode=0 Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.431314 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerDied","Data":"3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.438276 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.450965 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.466397 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.478615 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.493600 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.519697 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.528597 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.528645 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.528659 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.528678 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.528690 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.534251 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.554765 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.568329 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.582328 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.594492 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.609928 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.624313 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.631726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.631762 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.631773 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.631789 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.631800 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.639941 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.654739 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.667139 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:36Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.735358 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.735397 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.735407 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.735422 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.735437 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.838483 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.838649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.838717 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.838782 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.838851 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.941278 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.941468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.941543 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.941647 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:36 crc kubenswrapper[4807]: I1205 12:06:36.941704 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:36Z","lastTransitionTime":"2025-12-05T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.044809 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.045003 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.045059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.045117 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.045193 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.147391 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.147624 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.147684 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.147741 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.147802 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.235388 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.235610 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.236056 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.236115 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.237487 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.237762 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.255580 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.255848 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.255909 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.255985 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.256042 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.359788 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.359829 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.359838 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.359851 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.359860 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.454325 4807 generic.go:334] "Generic (PLEG): container finished" podID="f23809b6-3d63-484f-8e7c-9dc5dac87ebd" containerID="1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4" exitCode=0 Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.454832 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerDied","Data":"1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.462541 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.462595 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.462608 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.462626 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.462637 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.472875 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.489592 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.507172 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.520906 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.521090 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.521154 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:45.521137292 +0000 UTC m=+35.015000571 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.529279 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.542880 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.556342 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.564700 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.564744 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.564753 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.564767 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.564779 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.569911 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.594393 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.605545 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.615045 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.621631 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.621763 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.621815 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:06:45.621788571 +0000 UTC m=+35.115651830 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.621889 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.621898 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.621909 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.621922 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.621966 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:45.621950994 +0000 UTC m=+35.115814323 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.622259 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.622291 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.622307 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.622330 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.622333 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.622376 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:45.622368324 +0000 UTC m=+35.116231593 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:37 crc kubenswrapper[4807]: E1205 12:06:37.622388 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:45.622383464 +0000 UTC m=+35.116246733 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.625631 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.638494 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.648959 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.667042 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.667971 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.668014 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.668023 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.668038 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.668048 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.687184 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:37Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.770835 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.770899 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.770915 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.770938 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.770954 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.873077 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.873147 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.873171 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.873203 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.873225 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.977388 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.977451 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.977462 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.977480 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:37 crc kubenswrapper[4807]: I1205 12:06:37.977492 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:37Z","lastTransitionTime":"2025-12-05T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.080332 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.080390 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.080408 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.080431 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.080448 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.183198 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.183999 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.184330 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.184468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.184613 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.287669 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.288808 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.288948 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.289080 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.289205 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.396789 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.396840 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.396851 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.396869 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.396878 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.459020 4807 generic.go:334] "Generic (PLEG): container finished" podID="f23809b6-3d63-484f-8e7c-9dc5dac87ebd" containerID="2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db" exitCode=0 Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.459088 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerDied","Data":"2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.464330 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.464624 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.464659 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.471977 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.483467 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.489377 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.494659 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.500330 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.500393 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.500405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.500423 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.500435 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.505653 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.523300 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.536741 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.557163 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.571559 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.584639 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.597846 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.603353 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.603388 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.603395 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.603410 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.603421 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.612011 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.626452 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.666450 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.692906 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.705972 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.706018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.706029 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.706044 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.706056 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.708476 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.718062 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.729252 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.742060 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.756090 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.768930 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.788224 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.808096 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.808490 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.808577 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.808589 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.808609 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.808622 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.821497 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.836324 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.853375 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.867213 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.882440 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.895476 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.908454 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.911112 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.911177 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.911190 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.911206 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.911218 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:38Z","lastTransitionTime":"2025-12-05T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:38 crc kubenswrapper[4807]: I1205 12:06:38.922832 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:38Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.014048 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.014087 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.014095 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.014110 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.014119 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.117841 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.117901 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.117918 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.117944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.117961 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.220834 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.220883 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.220892 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.220908 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.220920 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.235452 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.235453 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.235692 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:39 crc kubenswrapper[4807]: E1205 12:06:39.235703 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:39 crc kubenswrapper[4807]: E1205 12:06:39.235615 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:39 crc kubenswrapper[4807]: E1205 12:06:39.235820 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.323448 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.323492 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.323508 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.323554 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.323569 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.427266 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.427331 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.427346 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.427367 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.427380 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.474380 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" event={"ID":"f23809b6-3d63-484f-8e7c-9dc5dac87ebd","Type":"ContainerStarted","Data":"2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.474566 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.493360 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.511955 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.530814 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.530858 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.530870 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.530888 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.530903 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.530710 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.545406 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.548218 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.561097 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.573657 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.588893 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.604017 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.617010 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.628112 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.633040 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.633072 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.633085 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.633103 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.633113 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.642234 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.653014 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.672933 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.686189 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.704633 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.716215 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.735226 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.735270 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.735309 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.735327 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.735338 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.738400 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.763848 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.777147 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.789814 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.806701 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.834238 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.837941 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.838177 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.838265 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.838343 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.838430 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.848567 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.864271 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.879952 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.896969 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.910188 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.924121 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.941124 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.941620 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.941671 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.941686 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.941705 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.941742 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:39Z","lastTransitionTime":"2025-12-05T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:39 crc kubenswrapper[4807]: I1205 12:06:39.954874 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:39Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.044899 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.044956 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.044968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.044988 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.045001 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.147330 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.147377 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.147390 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.147405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.147416 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.249766 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.249804 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.249814 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.249831 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.249841 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.352188 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.352243 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.352256 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.352278 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.352291 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.455011 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.455064 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.455078 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.455098 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.455112 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.558148 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.558201 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.558217 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.558237 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.558255 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.660623 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.660657 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.660667 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.660685 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.660698 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.764099 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.764869 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.764932 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.764965 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.764989 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.868410 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.868448 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.868479 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.868493 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.868510 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.972135 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.972176 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.972188 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.972205 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:40 crc kubenswrapper[4807]: I1205 12:06:40.972217 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:40Z","lastTransitionTime":"2025-12-05T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.074511 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.074605 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.074619 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.074639 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.074652 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.178120 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.178177 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.178188 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.178206 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.178219 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.234824 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.234896 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.234936 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:41 crc kubenswrapper[4807]: E1205 12:06:41.234971 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:41 crc kubenswrapper[4807]: E1205 12:06:41.235091 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:41 crc kubenswrapper[4807]: E1205 12:06:41.235180 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.250996 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.267926 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.280702 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.280752 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.280762 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.280780 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.280792 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.289286 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.306285 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.320747 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.334882 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.346695 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.358094 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.376925 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.383247 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.383288 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.383296 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.383311 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.383320 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.396818 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.410546 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.427660 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.441127 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.457037 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.473158 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:41Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.485616 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.485667 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.485679 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.485697 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.485708 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.588447 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.588486 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.588495 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.588508 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.588516 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.691250 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.691317 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.691328 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.691346 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.691359 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.803601 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.803684 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.803706 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.803737 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.803766 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.905683 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.905722 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.905731 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.905744 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:41 crc kubenswrapper[4807]: I1205 12:06:41.905754 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:41Z","lastTransitionTime":"2025-12-05T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.008880 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.009188 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.009201 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.009216 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.009227 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.111468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.111575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.111590 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.111606 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.111633 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.214647 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.214700 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.214721 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.214750 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.214773 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.317056 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.317133 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.317153 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.317179 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.317198 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.420503 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.420582 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.420595 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.420615 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.420630 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.487037 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/0.log" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.491809 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684" exitCode=1 Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.491883 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.493061 4807 scope.go:117] "RemoveContainer" containerID="18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.523467 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.523513 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.523555 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.523576 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.523591 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.530659 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.544685 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.564660 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:41Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 12:06:40.777880 6118 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 12:06:40.777970 6118 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 12:06:40.777996 6118 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 12:06:40.778031 6118 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 12:06:40.778045 6118 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 12:06:40.778082 6118 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 12:06:40.778098 6118 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 12:06:40.778105 6118 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 12:06:40.778113 6118 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 12:06:40.778120 6118 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 12:06:40.778350 6118 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 12:06:40.778379 6118 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 12:06:40.778407 6118 factory.go:656] Stopping watch factory\\\\nI1205 12:06:40.778427 6118 ovnkube.go:599] Stopped ovnkube\\\\nI1205 12:06:40.778472 6118 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.581556 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.593709 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.608847 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.621479 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.625690 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.625846 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.625958 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.626064 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.626153 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.634982 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.649424 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.663424 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.677798 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.692292 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.702608 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.713678 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.726123 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:42Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.728924 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.728955 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.728966 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.728980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.728989 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.831934 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.832002 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.832014 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.832058 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.832071 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.936051 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.936107 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.936119 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.936135 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:42 crc kubenswrapper[4807]: I1205 12:06:42.936146 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:42Z","lastTransitionTime":"2025-12-05T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.038843 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.038887 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.038901 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.038919 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.038930 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.141471 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.141506 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.141515 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.141552 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.141562 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.235126 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.235190 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.235329 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.235610 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.235697 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.235501 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.251440 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.251498 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.251512 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.251563 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.251589 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.267705 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.267759 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.267768 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.267787 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.267797 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.281746 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.287629 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.287677 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.287689 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.287708 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.287723 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.302563 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.307593 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.307658 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.307672 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.307690 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.307703 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.323598 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.328893 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.328936 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.328945 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.328960 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.328971 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.334351 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f"] Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.334835 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.337284 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.337430 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.348557 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.349276 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.354962 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.355006 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.355026 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.355051 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.355068 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.365684 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.368736 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: E1205 12:06:43.368943 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.371054 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.371119 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.371133 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.371151 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.371164 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.379683 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.396500 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.415399 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.431905 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.456209 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:41Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 12:06:40.777880 6118 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 12:06:40.777970 6118 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 12:06:40.777996 6118 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 12:06:40.778031 6118 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 12:06:40.778045 6118 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 12:06:40.778082 6118 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 12:06:40.778098 6118 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 12:06:40.778105 6118 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 12:06:40.778113 6118 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 12:06:40.778120 6118 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 12:06:40.778350 6118 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 12:06:40.778379 6118 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 12:06:40.778407 6118 factory.go:656] Stopping watch factory\\\\nI1205 12:06:40.778427 6118 ovnkube.go:599] Stopped ovnkube\\\\nI1205 12:06:40.778472 6118 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.474242 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.474281 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.474292 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.474310 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.474319 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.484707 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6eaedc27-e308-43e8-b3a9-14d094f179c2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.484775 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6eaedc27-e308-43e8-b3a9-14d094f179c2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.484826 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6eaedc27-e308-43e8-b3a9-14d094f179c2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.484867 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdrrl\" (UniqueName: \"kubernetes.io/projected/6eaedc27-e308-43e8-b3a9-14d094f179c2-kube-api-access-zdrrl\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.486595 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.498950 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/0.log" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.502044 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.502573 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.504767 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.522225 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.539182 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.556790 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.574755 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.576638 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.576691 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.576704 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.576718 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.576729 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.585850 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6eaedc27-e308-43e8-b3a9-14d094f179c2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.585929 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6eaedc27-e308-43e8-b3a9-14d094f179c2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.585960 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6eaedc27-e308-43e8-b3a9-14d094f179c2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.585991 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdrrl\" (UniqueName: \"kubernetes.io/projected/6eaedc27-e308-43e8-b3a9-14d094f179c2-kube-api-access-zdrrl\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.587179 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6eaedc27-e308-43e8-b3a9-14d094f179c2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.587352 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6eaedc27-e308-43e8-b3a9-14d094f179c2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.592123 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.592669 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6eaedc27-e308-43e8-b3a9-14d094f179c2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.608182 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdrrl\" (UniqueName: \"kubernetes.io/projected/6eaedc27-e308-43e8-b3a9-14d094f179c2-kube-api-access-zdrrl\") pod \"ovnkube-control-plane-749d76644c-2n72f\" (UID: \"6eaedc27-e308-43e8-b3a9-14d094f179c2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.611383 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.630920 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.646131 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.650875 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.664073 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.680274 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.680720 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.680810 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.680929 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.681085 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.681242 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.693233 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.712408 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.724060 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.745185 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:41Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 12:06:40.777880 6118 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 12:06:40.777970 6118 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 12:06:40.777996 6118 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 12:06:40.778031 6118 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 12:06:40.778045 6118 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 12:06:40.778082 6118 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 12:06:40.778098 6118 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 12:06:40.778105 6118 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 12:06:40.778113 6118 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 12:06:40.778120 6118 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 12:06:40.778350 6118 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 12:06:40.778379 6118 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 12:06:40.778407 6118 factory.go:656] Stopping watch factory\\\\nI1205 12:06:40.778427 6118 ovnkube.go:599] Stopped ovnkube\\\\nI1205 12:06:40.778472 6118 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.757617 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.772082 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.785044 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.785100 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.785117 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.785140 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.785157 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.789120 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.805124 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.820727 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.834680 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.852322 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.866790 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.880804 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:43Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.887368 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.887404 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.887415 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.887433 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.887445 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.989986 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.990036 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.990046 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.990060 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:43 crc kubenswrapper[4807]: I1205 12:06:43.990068 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:43Z","lastTransitionTime":"2025-12-05T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: W1205 12:06:44.086863 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6eaedc27_e308_43e8_b3a9_14d094f179c2.slice/crio-2810a8bd2af32c87b27ea2b9fdfde92b2bb241938097bff0f3badb1a69386113 WatchSource:0}: Error finding container 2810a8bd2af32c87b27ea2b9fdfde92b2bb241938097bff0f3badb1a69386113: Status 404 returned error can't find the container with id 2810a8bd2af32c87b27ea2b9fdfde92b2bb241938097bff0f3badb1a69386113 Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.092063 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.092137 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.092152 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.092178 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.092201 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.194374 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.194408 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.194418 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.194445 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.194456 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.265608 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.280169 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.293153 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.296851 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.296880 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.296891 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.296908 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.296919 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.308990 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.322970 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.350936 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.398845 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.400056 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.400115 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.400126 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.400147 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.400157 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.419911 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.437618 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.458462 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.485546 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.501884 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.503013 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.503054 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.503073 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.503090 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.503102 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.507772 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/1.log" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.508438 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/0.log" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.515605 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a" exitCode=1 Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.515734 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.515871 4807 scope.go:117] "RemoveContainer" containerID="18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.516496 4807 scope.go:117] "RemoveContainer" containerID="cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a" Dec 05 12:06:44 crc kubenswrapper[4807]: E1205 12:06:44.516732 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.517198 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" event={"ID":"6eaedc27-e308-43e8-b3a9-14d094f179c2","Type":"ContainerStarted","Data":"6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.517240 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" event={"ID":"6eaedc27-e308-43e8-b3a9-14d094f179c2","Type":"ContainerStarted","Data":"2810a8bd2af32c87b27ea2b9fdfde92b2bb241938097bff0f3badb1a69386113"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.525747 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:41Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 12:06:40.777880 6118 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 12:06:40.777970 6118 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 12:06:40.777996 6118 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 12:06:40.778031 6118 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 12:06:40.778045 6118 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 12:06:40.778082 6118 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 12:06:40.778098 6118 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 12:06:40.778105 6118 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 12:06:40.778113 6118 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 12:06:40.778120 6118 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 12:06:40.778350 6118 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 12:06:40.778379 6118 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 12:06:40.778407 6118 factory.go:656] Stopping watch factory\\\\nI1205 12:06:40.778427 6118 ovnkube.go:599] Stopped ovnkube\\\\nI1205 12:06:40.778472 6118 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.541012 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.555299 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.571455 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.584159 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.604357 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.606181 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.606257 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.606271 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.606294 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.606307 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.619632 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.638763 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:41Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 12:06:40.777880 6118 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 12:06:40.777970 6118 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 12:06:40.777996 6118 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 12:06:40.778031 6118 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 12:06:40.778045 6118 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 12:06:40.778082 6118 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 12:06:40.778098 6118 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 12:06:40.778105 6118 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 12:06:40.778113 6118 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 12:06:40.778120 6118 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 12:06:40.778350 6118 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 12:06:40.778379 6118 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 12:06:40.778407 6118 factory.go:656] Stopping watch factory\\\\nI1205 12:06:40.778427 6118 ovnkube.go:599] Stopped ovnkube\\\\nI1205 12:06:40.778472 6118 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.655434 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.676737 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.697494 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.708746 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.708804 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.708816 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.708838 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.708852 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.712393 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.729792 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.744953 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.761038 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.774795 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.789412 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.804833 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.811594 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.811639 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.811649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.811666 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.811677 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.816266 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.820077 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-d4t4l"] Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.820596 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:44 crc kubenswrapper[4807]: E1205 12:06:44.820665 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.826222 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.839307 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.853643 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.866115 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.876602 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.888497 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.897848 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.897920 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll4br\" (UniqueName: \"kubernetes.io/projected/a0fb825b-4e77-44f1-a14b-5a3660dd1799-kube-api-access-ll4br\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.902903 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.915036 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.915104 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.915117 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.915138 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.915150 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:44Z","lastTransitionTime":"2025-12-05T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.919410 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.931588 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.944047 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.957380 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.970670 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.997716 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:44Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.998586 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:44 crc kubenswrapper[4807]: I1205 12:06:44.998636 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll4br\" (UniqueName: \"kubernetes.io/projected/a0fb825b-4e77-44f1-a14b-5a3660dd1799-kube-api-access-ll4br\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:44 crc kubenswrapper[4807]: E1205 12:06:44.998878 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:44 crc kubenswrapper[4807]: E1205 12:06:44.999046 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:45.498999117 +0000 UTC m=+34.992862386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.012805 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.019318 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.019596 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.019663 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.019726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.019796 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.023573 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll4br\" (UniqueName: \"kubernetes.io/projected/a0fb825b-4e77-44f1-a14b-5a3660dd1799-kube-api-access-ll4br\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.035432 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:41Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 12:06:40.777880 6118 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 12:06:40.777970 6118 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 12:06:40.777996 6118 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 12:06:40.778031 6118 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 12:06:40.778045 6118 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 12:06:40.778082 6118 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 12:06:40.778098 6118 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 12:06:40.778105 6118 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 12:06:40.778113 6118 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 12:06:40.778120 6118 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 12:06:40.778350 6118 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 12:06:40.778379 6118 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 12:06:40.778407 6118 factory.go:656] Stopping watch factory\\\\nI1205 12:06:40.778427 6118 ovnkube.go:599] Stopped ovnkube\\\\nI1205 12:06:40.778472 6118 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.053249 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.069389 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.085060 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.104373 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.122562 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.122915 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.122997 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.123104 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.123190 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.226123 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.226568 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.226768 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.227114 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.227199 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.235392 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.235467 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.235476 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.236052 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.236005 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.236211 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.331019 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.331184 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.331268 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.331334 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.331391 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.435062 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.435479 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.435618 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.435708 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.435792 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.502859 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.503106 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.503176 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:46.50315888 +0000 UTC m=+35.997022159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.522332 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" event={"ID":"6eaedc27-e308-43e8-b3a9-14d094f179c2","Type":"ContainerStarted","Data":"5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.524422 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/1.log" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.529059 4807 scope.go:117] "RemoveContainer" containerID="cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.529227 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.538285 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.538321 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.538459 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.538471 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.538488 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.538500 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.549690 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.561094 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.574211 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.589137 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.604137 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.604511 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.604608 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:01.604588456 +0000 UTC m=+51.098451775 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.611026 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.625748 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.640928 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.640976 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.640987 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.641002 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.641012 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.650546 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a898c31dc99c40e15cf152c3b30827985a258f6dff71bee77b41cb81b0e684\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:41Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 12:06:40.777880 6118 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 12:06:40.777970 6118 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 12:06:40.777996 6118 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 12:06:40.778031 6118 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 12:06:40.778045 6118 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 12:06:40.778082 6118 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 12:06:40.778098 6118 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 12:06:40.778105 6118 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 12:06:40.778113 6118 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 12:06:40.778120 6118 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 12:06:40.778350 6118 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 12:06:40.778379 6118 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 12:06:40.778407 6118 factory.go:656] Stopping watch factory\\\\nI1205 12:06:40.778427 6118 ovnkube.go:599] Stopped ovnkube\\\\nI1205 12:06:40.778472 6118 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.665576 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.677025 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.692219 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.706334 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.706470 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.706502 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:07:01.706465313 +0000 UTC m=+51.200328582 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.706586 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.706642 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.706650 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:01.706632487 +0000 UTC m=+51.200495946 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.706701 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.706907 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.706962 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.706978 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.707044 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:01.707023087 +0000 UTC m=+51.200886356 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.707429 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.707477 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.707498 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:45 crc kubenswrapper[4807]: E1205 12:06:45.711556 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:01.711508569 +0000 UTC m=+51.205371838 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.713591 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.726149 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.740252 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.743830 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.743871 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.743879 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.743895 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.743906 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.752997 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.765507 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.777066 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.789355 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.803448 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.814954 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.828715 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.840158 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.846181 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.846233 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.846249 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.846284 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.846325 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.862021 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.875911 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.902235 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.919694 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.935684 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.949179 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.949243 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.949255 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.949284 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.949299 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:45Z","lastTransitionTime":"2025-12-05T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.952474 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.965448 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.982088 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:45 crc kubenswrapper[4807]: I1205 12:06:45.993766 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:45Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.006561 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:46Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.019460 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:46Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.034056 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:46Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.052266 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.052310 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.052320 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.052338 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.052359 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.155330 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.155379 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.155393 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.155411 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.155425 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.234509 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:46 crc kubenswrapper[4807]: E1205 12:06:46.234720 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.257109 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.257151 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.257162 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.257176 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.257186 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.359175 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.359243 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.359258 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.359281 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.359299 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.461334 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.461385 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.461396 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.461411 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.461423 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.514174 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:46 crc kubenswrapper[4807]: E1205 12:06:46.514347 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:46 crc kubenswrapper[4807]: E1205 12:06:46.514421 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:48.514401064 +0000 UTC m=+38.008264333 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.563516 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.563571 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.563583 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.563596 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.563606 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.666132 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.666180 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.666192 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.666209 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.666220 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.768900 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.768936 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.768945 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.768958 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.768967 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.871580 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.871626 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.871638 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.871658 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.871671 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.973944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.974009 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.974019 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.974034 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:46 crc kubenswrapper[4807]: I1205 12:06:46.974044 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:46Z","lastTransitionTime":"2025-12-05T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.077373 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.077450 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.077468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.077493 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.077511 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.180506 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.180614 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.180633 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.180648 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.180658 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.234656 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.234698 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:47 crc kubenswrapper[4807]: E1205 12:06:47.234812 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.234828 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:47 crc kubenswrapper[4807]: E1205 12:06:47.234954 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:47 crc kubenswrapper[4807]: E1205 12:06:47.235043 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.283749 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.283823 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.283839 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.283860 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.283873 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.386965 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.387038 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.387056 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.387082 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.387099 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.490009 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.490060 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.490072 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.490089 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.490102 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.592626 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.592674 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.592685 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.592705 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.592717 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.695775 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.695856 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.695869 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.695887 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.695925 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.799178 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.799265 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.799295 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.799325 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.799348 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.902522 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.902584 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.902595 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.902611 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:47 crc kubenswrapper[4807]: I1205 12:06:47.902622 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:47Z","lastTransitionTime":"2025-12-05T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.005930 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.005997 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.006018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.006043 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.006063 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.108343 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.108403 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.108414 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.108428 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.108438 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.211271 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.211323 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.211332 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.211349 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.211362 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.235293 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:48 crc kubenswrapper[4807]: E1205 12:06:48.235571 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.314124 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.314206 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.314222 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.314238 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.314268 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.416511 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.416588 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.416599 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.416620 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.416632 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.518707 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.518747 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.518757 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.518771 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.518780 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.532863 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:48 crc kubenswrapper[4807]: E1205 12:06:48.533046 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:48 crc kubenswrapper[4807]: E1205 12:06:48.533123 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:06:52.533106265 +0000 UTC m=+42.026969534 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.621759 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.621814 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.621823 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.621839 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.621849 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.724020 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.724055 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.724068 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.724083 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.724095 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.827222 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.827781 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.827797 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.827826 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.827846 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.931306 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.931376 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.931390 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.931415 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:48 crc kubenswrapper[4807]: I1205 12:06:48.931429 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:48Z","lastTransitionTime":"2025-12-05T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.034679 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.034744 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.034755 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.034782 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.034798 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.137447 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.137591 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.137652 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.137671 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.137685 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.234628 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.234720 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:49 crc kubenswrapper[4807]: E1205 12:06:49.234815 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.234947 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:49 crc kubenswrapper[4807]: E1205 12:06:49.235047 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:49 crc kubenswrapper[4807]: E1205 12:06:49.235181 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.239587 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.239629 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.239649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.239673 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.239690 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.342734 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.342786 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.342798 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.342816 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.342827 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.446647 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.446734 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.446745 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.446891 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.446905 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.554019 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.554345 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.554444 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.554576 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.554656 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.656860 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.656944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.656955 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.656980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.656994 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.759361 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.759442 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.759453 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.759475 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.759491 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.862724 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.862780 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.862791 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.862815 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.862827 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.965441 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.965493 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.965504 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.965545 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:49 crc kubenswrapper[4807]: I1205 12:06:49.965557 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:49Z","lastTransitionTime":"2025-12-05T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.068612 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.068671 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.068681 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.068699 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.068717 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.171379 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.171428 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.171437 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.171453 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.171464 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.235213 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:50 crc kubenswrapper[4807]: E1205 12:06:50.235417 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.274309 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.274353 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.274369 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.274389 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.274399 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.377387 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.377444 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.377457 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.377477 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.377490 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.480996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.481056 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.481066 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.481082 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.481092 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.583741 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.584038 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.584121 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.584205 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.584279 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.687153 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.687669 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.687762 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.687825 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.687894 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.791883 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.792503 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.792761 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.792990 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.793396 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.896680 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.896746 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.896758 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.896776 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:50 crc kubenswrapper[4807]: I1205 12:06:50.896788 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:50Z","lastTransitionTime":"2025-12-05T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.000275 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.000327 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.000338 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.000355 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.000366 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.103106 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.103494 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.103655 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.103792 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.103929 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.206118 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.206364 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.206443 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.206575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.206641 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.234634 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.234951 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:51 crc kubenswrapper[4807]: E1205 12:06:51.235023 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:51 crc kubenswrapper[4807]: E1205 12:06:51.235069 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.235054 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:51 crc kubenswrapper[4807]: E1205 12:06:51.235138 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.255003 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.271678 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.286817 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.298694 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.309273 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.309324 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.309334 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.309361 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.309372 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.313430 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.331420 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.344320 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.355216 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.370830 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.389379 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.411807 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.411881 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.411892 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.411910 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.411602 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.411922 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.426946 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.447228 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.461101 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.473664 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.490831 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.511906 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:51Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.514557 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.514599 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.514615 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.514630 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.514639 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.617153 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.617201 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.617209 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.617223 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.617233 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.719487 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.719542 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.719552 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.719565 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.719576 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.821951 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.822013 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.822031 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.822052 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.822065 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.923864 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.923897 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.923905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.923918 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:51 crc kubenswrapper[4807]: I1205 12:06:51.923929 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:51Z","lastTransitionTime":"2025-12-05T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.026111 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.026155 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.026167 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.026183 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.026194 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.129106 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.129183 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.129197 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.129215 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.129229 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.232121 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.232160 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.232176 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.232192 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.232203 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.234641 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:52 crc kubenswrapper[4807]: E1205 12:06:52.234772 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.335554 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.335612 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.335628 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.335651 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.335677 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.439304 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.439807 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.439972 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.440055 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.440118 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.543359 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.543664 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.543757 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.543837 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.543907 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.578349 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:52 crc kubenswrapper[4807]: E1205 12:06:52.578573 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:52 crc kubenswrapper[4807]: E1205 12:06:52.578680 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:00.578654984 +0000 UTC m=+50.072518333 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.646626 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.646857 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.646958 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.647065 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.647158 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.750906 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.750957 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.750968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.750989 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.751000 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.853911 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.853963 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.853976 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.853996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.854009 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.956662 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.956724 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.956740 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.956758 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:52 crc kubenswrapper[4807]: I1205 12:06:52.956772 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:52Z","lastTransitionTime":"2025-12-05T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.059462 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.059511 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.059538 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.059556 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.059568 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.161926 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.161980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.161989 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.162004 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.162016 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.235191 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.235269 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.235359 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.235445 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.235622 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.236251 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.265333 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.265456 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.265468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.265490 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.265541 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.368414 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.368485 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.368505 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.368547 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.368565 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.394666 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.394723 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.394737 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.394754 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.394769 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.408010 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:53Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.411862 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.411933 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.411944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.411967 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.411978 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.425413 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:53Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.428921 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.428972 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.428987 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.429008 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.429020 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.440609 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:53Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.443822 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.443850 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.443858 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.443870 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.443880 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.461739 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:53Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.466408 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.466437 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.466447 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.466461 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.466471 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.483547 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:53Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:53 crc kubenswrapper[4807]: E1205 12:06:53.483663 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.494859 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.494933 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.494953 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.494977 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.494992 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.597592 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.597663 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.597678 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.597704 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.597717 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.700130 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.700190 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.700202 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.700223 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.700237 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.803669 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.803734 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.803751 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.803776 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.803793 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.906546 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.906600 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.906611 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.906627 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:53 crc kubenswrapper[4807]: I1205 12:06:53.906639 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:53Z","lastTransitionTime":"2025-12-05T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.009455 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.009503 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.009513 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.009551 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.009572 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.112789 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.112845 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.112856 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.112874 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.112888 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.215652 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.215758 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.215773 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.215791 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.215805 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.235182 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:54 crc kubenswrapper[4807]: E1205 12:06:54.235333 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.319017 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.319075 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.319092 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.319111 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.319124 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.421837 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.421905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.421919 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.421941 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.421956 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.525050 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.525330 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.525406 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.525425 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.525438 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.631204 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.631277 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.631291 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.631310 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.631323 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.734543 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.734601 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.734612 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.734627 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.734637 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.837970 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.838018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.838044 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.838063 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.838082 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.940457 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.940499 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.940508 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.940568 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:54 crc kubenswrapper[4807]: I1205 12:06:54.940580 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:54Z","lastTransitionTime":"2025-12-05T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.044944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.044985 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.044998 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.045018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.045030 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.147758 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.147818 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.147829 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.147847 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.147859 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.235076 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.235121 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:55 crc kubenswrapper[4807]: E1205 12:06:55.235250 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.235269 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:55 crc kubenswrapper[4807]: E1205 12:06:55.235385 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:55 crc kubenswrapper[4807]: E1205 12:06:55.235467 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.250578 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.250742 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.250796 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.250823 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.250843 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.353810 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.353864 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.353874 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.353890 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.353901 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.456149 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.456215 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.456232 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.456256 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.456273 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.560308 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.560357 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.560367 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.560385 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.560395 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.663590 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.663651 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.663662 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.663682 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.663695 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.766196 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.766260 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.766272 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.766291 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.766306 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.870129 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.870166 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.870184 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.870201 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.870213 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.972875 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.972932 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.972941 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.972959 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:55 crc kubenswrapper[4807]: I1205 12:06:55.972971 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:55Z","lastTransitionTime":"2025-12-05T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.075845 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.075905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.075917 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.075934 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.075948 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.180128 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.180177 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.180186 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.180202 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.180215 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.234890 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:56 crc kubenswrapper[4807]: E1205 12:06:56.235121 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.282729 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.282790 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.282800 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.282819 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.282833 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.385830 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.385890 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.385908 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.385930 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.385943 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.489019 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.489068 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.489079 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.489095 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.489105 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.592475 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.592609 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.592632 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.592658 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.592675 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.695560 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.695598 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.695607 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.695622 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.695635 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.798817 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.798879 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.798896 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.798918 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.798929 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.902717 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.902771 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.902781 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.902805 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:56 crc kubenswrapper[4807]: I1205 12:06:56.902823 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:56Z","lastTransitionTime":"2025-12-05T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.005872 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.005921 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.005930 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.005946 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.005958 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.108901 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.108980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.108994 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.109017 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.109035 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.217590 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.217672 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.217688 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.217713 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.217728 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.235176 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.235308 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.235364 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:57 crc kubenswrapper[4807]: E1205 12:06:57.235488 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:57 crc kubenswrapper[4807]: E1205 12:06:57.235632 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:57 crc kubenswrapper[4807]: E1205 12:06:57.235776 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.321123 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.321201 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.321214 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.321237 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.321255 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.424296 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.424376 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.424387 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.424405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.424420 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.527174 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.527242 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.527253 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.527278 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.527290 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.630243 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.630299 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.630309 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.630330 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.630344 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.733594 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.733649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.733667 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.733692 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.733710 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.836553 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.836606 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.836619 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.836640 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.836653 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.939075 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.939662 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.939740 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.939811 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:57 crc kubenswrapper[4807]: I1205 12:06:57.939880 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:57Z","lastTransitionTime":"2025-12-05T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.046008 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.046052 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.046063 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.046080 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.046092 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.149776 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.149892 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.149914 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.149971 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.149989 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.235027 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:06:58 crc kubenswrapper[4807]: E1205 12:06:58.235615 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.252575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.252966 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.253101 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.253201 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.253304 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.356573 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.356613 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.356775 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.356795 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.356807 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.459985 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.460030 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.460042 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.460058 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.460068 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.562919 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.562976 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.562995 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.563018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.563036 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.666372 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.666905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.667091 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.667284 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.667451 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.771343 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.771442 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.771462 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.771485 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.771502 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.874793 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.875089 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.875209 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.875301 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.875411 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.977888 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.977977 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.977992 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.978013 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:58 crc kubenswrapper[4807]: I1205 12:06:58.978025 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:58Z","lastTransitionTime":"2025-12-05T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.080861 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.080928 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.080944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.080966 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.080980 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.184151 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.184228 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.184252 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.184280 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.184303 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.235077 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.235159 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:06:59 crc kubenswrapper[4807]: E1205 12:06:59.235288 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:06:59 crc kubenswrapper[4807]: E1205 12:06:59.235855 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.236020 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:06:59 crc kubenswrapper[4807]: E1205 12:06:59.236128 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.236508 4807 scope.go:117] "RemoveContainer" containerID="cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.286697 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.287088 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.287100 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.287117 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.287126 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.390081 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.390131 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.390142 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.390166 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.390180 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.493726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.493775 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.493786 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.493805 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.493815 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.575648 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/1.log" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.579304 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.579889 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.596844 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.596901 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.596910 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.596932 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.596942 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.598361 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.613857 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.627517 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.647193 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.668360 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.690804 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.699861 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.700331 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.700491 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.700646 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.700758 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.715958 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.735438 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.749059 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.763922 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.775821 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.786625 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.798063 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.802568 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.802607 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.802618 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.802634 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.802645 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.809902 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.832796 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.847231 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.870833 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:06:59Z is after 2025-08-24T17:21:41Z" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.905461 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.905498 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.905507 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.905520 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:06:59 crc kubenswrapper[4807]: I1205 12:06:59.905544 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:06:59Z","lastTransitionTime":"2025-12-05T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.007844 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.007889 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.007903 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.007921 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.007932 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.111578 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.111614 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.111622 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.111640 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.111652 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.250262 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:00 crc kubenswrapper[4807]: E1205 12:07:00.250831 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.253560 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.253599 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.253609 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.253624 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.253636 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.356206 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.356252 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.356266 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.356288 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.356304 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.458890 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.458929 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.458939 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.458953 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.458964 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.561209 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.561241 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.561250 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.561263 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.561272 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.583375 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/2.log" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.583913 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/1.log" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.586214 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f" exitCode=1 Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.586260 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.586306 4807 scope.go:117] "RemoveContainer" containerID="cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.586848 4807 scope.go:117] "RemoveContainer" containerID="88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f" Dec 05 12:07:00 crc kubenswrapper[4807]: E1205 12:07:00.586997 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.592731 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:00 crc kubenswrapper[4807]: E1205 12:07:00.592922 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:07:00 crc kubenswrapper[4807]: E1205 12:07:00.592982 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:16.592965129 +0000 UTC m=+66.086828398 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.602278 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.613093 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.629211 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.641737 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.656949 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.664007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.664126 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.664218 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.664311 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.664389 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.670275 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.683433 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.695826 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.705685 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.716337 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.726057 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.737390 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.750244 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.763033 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.766961 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.766991 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.767001 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.767014 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.767024 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.782825 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.793662 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.810702 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:00Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.870047 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.870093 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.870108 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.870126 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.870152 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.972828 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.972904 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.972928 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.972961 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:00 crc kubenswrapper[4807]: I1205 12:07:00.973005 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:00Z","lastTransitionTime":"2025-12-05T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.074858 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.074897 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.074908 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.074926 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.074938 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.177729 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.177782 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.177792 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.177808 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.177817 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.239626 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.239714 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.239793 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.239869 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.240210 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.240453 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.256123 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.274767 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9d5b8dab6252965ab9a1a0f5e70f8fcd7d8297ccc3b846681d80c6b7ce685a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\" {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.496979 6264 services_controller.go:444] Built service openshift-machine-api/machine-api-operator-machine-webhook LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496998 6264 services_controller.go:445] Built service openshift-machine-api/machine-api-operator-machine-webhook LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 12:06:43.496980 6264 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:06:43.497050 6264 services_controller.go:451] Built service openshift-machine-api/machine-api-operator-machine-webhook cluster-wide LB for network=de\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.280173 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.280220 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.280231 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.280247 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.280259 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.299515 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.315196 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.329457 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.345440 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.360498 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.378136 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.382724 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.382796 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.382808 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.382829 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.382841 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.393840 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.411947 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.427299 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.443453 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.455159 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.466239 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.479107 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.485376 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.485554 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.485620 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.485691 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.485750 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.495832 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.509123 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.591372 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.591414 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.591428 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.591455 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.591470 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.594731 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/2.log" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.599790 4807 scope.go:117] "RemoveContainer" containerID="88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f" Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.599972 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.614704 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.628315 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.645107 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.659848 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.677952 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.694575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.694614 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.694630 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.694651 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.694666 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.704349 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.704808 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.704972 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:33.704941394 +0000 UTC m=+83.198804693 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.714491 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.730148 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.751231 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.766952 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.786266 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.796407 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.796435 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.796445 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.796457 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.796467 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.799350 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.804984 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.805117 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.805164 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.805194 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805342 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805372 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805386 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805431 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:33.805415078 +0000 UTC m=+83.299278347 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805587 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805636 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805629 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:07:33.805591482 +0000 UTC m=+83.299454791 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805656 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805736 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:33.805722176 +0000 UTC m=+83.299585475 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805864 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: E1205 12:07:01.805983 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:33.805969571 +0000 UTC m=+83.299832840 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.814166 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.825968 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.840360 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.855189 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.871219 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.884900 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:01Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.898659 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.898712 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.898726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.898747 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:01 crc kubenswrapper[4807]: I1205 12:07:01.898762 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:01Z","lastTransitionTime":"2025-12-05T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.002018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.002077 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.002092 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.002109 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.002120 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.105107 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.105153 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.105166 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.105183 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.105195 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.208340 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.208411 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.208428 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.208450 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.208486 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.235005 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:02 crc kubenswrapper[4807]: E1205 12:07:02.235181 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.311560 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.311606 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.311618 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.311634 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.311646 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.414005 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.414048 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.414058 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.414072 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.414084 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.516962 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.517018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.517031 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.517053 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.517068 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.620178 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.620249 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.620260 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.620275 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.620286 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.723153 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.723211 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.723232 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.723257 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.723273 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.825740 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.825822 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.825840 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.825863 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.825875 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.929731 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.929799 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.929815 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.929835 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:02 crc kubenswrapper[4807]: I1205 12:07:02.929849 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:02Z","lastTransitionTime":"2025-12-05T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.032561 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.032653 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.032697 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.032720 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.032737 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.135564 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.135609 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.135621 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.135639 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.135652 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.235053 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.235086 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.235262 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.235294 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.235575 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.235805 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.239122 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.239259 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.239368 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.239505 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.239624 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.342128 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.342180 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.342195 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.342218 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.342234 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.444695 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.444768 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.444808 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.444841 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.444864 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.547913 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.547981 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.548004 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.548036 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.548062 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.608625 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.608690 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.608715 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.608742 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.608762 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.624079 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.629894 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.629955 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.629973 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.629996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.630014 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.650508 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.655091 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.655149 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.655167 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.655189 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.655201 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.672847 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.677955 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.678044 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.678066 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.678096 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.678122 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.698940 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.708361 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.708418 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.708441 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.708478 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.708504 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.728057 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:03 crc kubenswrapper[4807]: E1205 12:07:03.728455 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.730928 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.730973 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.730987 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.731007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.731022 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.833686 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.834129 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.834351 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.834585 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.834778 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.937976 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.938059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.938078 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.938102 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:03 crc kubenswrapper[4807]: I1205 12:07:03.938148 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:03Z","lastTransitionTime":"2025-12-05T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.041727 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.041779 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.041804 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.041833 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.041858 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.144148 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.144194 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.144208 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.144226 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.144239 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.234756 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:04 crc kubenswrapper[4807]: E1205 12:07:04.235007 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.247968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.248085 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.248108 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.248164 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.248188 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.351358 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.351395 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.351405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.351421 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.351431 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.453956 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.454013 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.454024 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.454038 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.454049 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.557423 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.557677 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.557773 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.557852 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.557921 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.660384 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.660431 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.660464 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.660484 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.660493 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.763238 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.763293 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.763306 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.763324 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.763336 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.866172 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.866224 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.866233 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.866254 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.866281 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.968900 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.968959 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.968967 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.968980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:04 crc kubenswrapper[4807]: I1205 12:07:04.968989 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:04Z","lastTransitionTime":"2025-12-05T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.072168 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.072245 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.072260 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.072280 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.072295 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.175948 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.176018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.176029 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.176045 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.176055 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.235000 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.235272 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.235030 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:05 crc kubenswrapper[4807]: E1205 12:07:05.235263 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:05 crc kubenswrapper[4807]: E1205 12:07:05.235424 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:05 crc kubenswrapper[4807]: E1205 12:07:05.235621 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.278135 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.278190 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.278206 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.278229 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.278250 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.381436 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.381490 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.381505 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.381555 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.381572 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.484339 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.484373 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.484383 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.484399 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.484410 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.587635 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.587684 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.587698 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.587718 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.587736 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.690584 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.690638 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.690649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.690665 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.690676 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.792775 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.792823 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.792837 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.792852 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.792864 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.899901 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.899950 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.900359 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.900429 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:05 crc kubenswrapper[4807]: I1205 12:07:05.900440 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:05Z","lastTransitionTime":"2025-12-05T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.002374 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.002418 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.002427 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.002442 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.002450 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.105252 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.105291 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.105301 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.105314 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.105323 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.208361 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.208403 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.208419 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.208441 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.208456 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.234798 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:06 crc kubenswrapper[4807]: E1205 12:07:06.235015 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.312368 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.312415 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.312425 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.312440 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.312450 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.415383 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.415489 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.415503 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.415543 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.415560 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.518741 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.518804 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.518818 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.518836 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.518849 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.621230 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.621273 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.621282 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.621300 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.621311 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.723824 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.723882 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.723893 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.723908 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.723918 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.827213 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.827263 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.827277 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.827298 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.827320 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.930713 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.930803 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.930826 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.930853 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:06 crc kubenswrapper[4807]: I1205 12:07:06.930872 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:06Z","lastTransitionTime":"2025-12-05T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.033116 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.033152 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.033161 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.033175 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.033184 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.061922 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.073854 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.084789 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.097870 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.119430 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.131902 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.135884 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.135942 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.135954 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.135975 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.135987 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.145586 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.158461 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.173760 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.189693 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.202503 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.219351 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.234481 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.234592 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.234481 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:07 crc kubenswrapper[4807]: E1205 12:07:07.234690 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:07 crc kubenswrapper[4807]: E1205 12:07:07.234836 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:07 crc kubenswrapper[4807]: E1205 12:07:07.234897 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.238508 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.238583 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.238595 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.238637 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.238651 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.242185 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.262690 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.278376 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.295504 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.310480 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.323844 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.335806 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.341683 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.341747 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.341760 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.341786 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.341799 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.444203 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.444338 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.444351 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.444369 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.444380 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.547730 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.547781 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.547792 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.547808 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.547819 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.650367 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.650429 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.650439 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.650451 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.650460 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.752915 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.752963 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.752977 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.752995 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.753006 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.856071 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.856120 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.856131 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.856146 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.856157 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.958895 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.958966 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.958983 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.959007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:07 crc kubenswrapper[4807]: I1205 12:07:07.959022 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:07Z","lastTransitionTime":"2025-12-05T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.061387 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.061434 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.061443 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.061456 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.061465 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.164173 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.164209 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.164218 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.164233 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.164243 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.234698 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:08 crc kubenswrapper[4807]: E1205 12:07:08.234854 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.267216 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.267293 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.267301 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.267333 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.267343 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.369732 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.369814 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.369827 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.369843 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.369855 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.472231 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.472268 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.472278 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.472290 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.472300 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.574685 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.574766 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.574784 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.574807 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.574823 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.676914 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.676954 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.676965 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.676982 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.676994 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.779996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.780043 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.780055 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.780073 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.780085 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.882638 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.882684 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.882696 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.882715 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.882726 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.985361 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.985405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.985415 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.985432 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:08 crc kubenswrapper[4807]: I1205 12:07:08.985444 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:08Z","lastTransitionTime":"2025-12-05T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.088222 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.088286 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.088319 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.088339 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.088352 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.191597 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.191675 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.191688 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.191708 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.191721 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.234469 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:09 crc kubenswrapper[4807]: E1205 12:07:09.234641 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.234631 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.234698 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:09 crc kubenswrapper[4807]: E1205 12:07:09.234852 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:09 crc kubenswrapper[4807]: E1205 12:07:09.234894 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.295311 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.295651 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.295800 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.295929 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.296073 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.399371 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.399433 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.399457 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.399484 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.399501 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.502066 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.502176 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.502195 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.502214 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.502226 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.605256 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.605287 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.605296 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.605309 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.605319 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.708003 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.708048 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.708058 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.708073 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.708082 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.809878 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.809917 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.809928 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.809944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.809955 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.912492 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.912575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.912593 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.912616 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:09 crc kubenswrapper[4807]: I1205 12:07:09.912632 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:09Z","lastTransitionTime":"2025-12-05T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.015741 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.015793 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.015810 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.015834 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.015851 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.118362 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.118412 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.118422 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.118444 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.118455 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.221659 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.221695 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.221705 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.221721 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.221733 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.234628 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:10 crc kubenswrapper[4807]: E1205 12:07:10.234783 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.324216 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.324252 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.324263 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.324281 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.324295 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.427453 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.427511 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.427562 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.427586 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.427606 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.531751 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.531802 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.531816 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.531839 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.531852 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.634363 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.634419 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.634429 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.634444 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.634454 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.737907 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.737994 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.738010 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.738030 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.738041 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.841952 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.841993 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.842002 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.842017 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.842026 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.944942 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.944997 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.945010 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.945027 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:10 crc kubenswrapper[4807]: I1205 12:07:10.945041 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:10Z","lastTransitionTime":"2025-12-05T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.048168 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.048477 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.048721 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.048920 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.049146 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.151848 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.151905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.151918 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.151937 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.151953 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.235079 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:11 crc kubenswrapper[4807]: E1205 12:07:11.235228 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.235297 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.235565 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:11 crc kubenswrapper[4807]: E1205 12:07:11.235566 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:11 crc kubenswrapper[4807]: E1205 12:07:11.235634 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.254314 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.254357 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.254369 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.254387 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.254400 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.256622 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.269921 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24431c1-0ace-441e-be42-4b199d46972b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.282938 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.296564 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.312151 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.329987 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.341872 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.351871 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.356410 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.356461 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.356492 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.356513 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.356547 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.363697 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.374794 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.387014 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.396261 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.405876 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.419197 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.432100 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.455357 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.459322 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.459358 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.459388 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.459405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.459416 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.469247 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.489595 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.562922 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.562959 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.562970 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.562984 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.562993 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.665400 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.665456 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.665465 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.665479 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.665490 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.767318 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.767364 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.767380 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.767400 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.767414 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.869106 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.869160 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.869172 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.869187 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.869198 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.972118 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.972149 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.972158 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.972171 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:11 crc kubenswrapper[4807]: I1205 12:07:11.972181 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:11Z","lastTransitionTime":"2025-12-05T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.075011 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.075371 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.075382 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.075396 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.075406 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.177728 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.177771 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.177781 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.177797 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.177806 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.235017 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:12 crc kubenswrapper[4807]: E1205 12:07:12.235217 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.280061 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.280105 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.280122 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.280144 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.280162 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.382405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.382448 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.382465 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.382481 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.382492 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.485045 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.485155 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.485165 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.485186 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.485198 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.588251 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.588301 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.588309 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.588323 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.588333 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.690378 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.690431 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.690443 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.690462 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.690473 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.792393 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.792476 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.792500 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.792574 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.792598 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.894726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.894966 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.894974 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.894990 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.895000 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.997461 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.997504 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.997517 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.997551 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:12 crc kubenswrapper[4807]: I1205 12:07:12.997565 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:12Z","lastTransitionTime":"2025-12-05T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.099857 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.099921 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.099933 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.099951 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.099961 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.202232 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.202276 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.202286 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.202303 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.202313 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.235192 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.235248 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.235340 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:13 crc kubenswrapper[4807]: E1205 12:07:13.235505 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:13 crc kubenswrapper[4807]: E1205 12:07:13.235639 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:13 crc kubenswrapper[4807]: E1205 12:07:13.235975 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.305166 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.305226 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.305236 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.305250 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.305260 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.407689 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.407737 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.407748 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.407765 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.407780 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.509931 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.510002 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.510014 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.510050 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.510067 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.612693 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.612741 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.612753 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.612771 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.612782 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.716166 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.717012 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.717048 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.717071 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.717085 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.820090 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.820128 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.820138 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.820153 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.820164 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.923242 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.923294 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.923309 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.923325 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.923338 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.988442 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.988571 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.988583 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.988600 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:13 crc kubenswrapper[4807]: I1205 12:07:13.988610 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:13Z","lastTransitionTime":"2025-12-05T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: E1205 12:07:14.008055 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.012863 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.012897 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.012908 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.012925 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.012942 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: E1205 12:07:14.034613 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.040023 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.040066 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.040080 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.040100 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.040116 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: E1205 12:07:14.053366 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.057713 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.057756 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.057765 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.057781 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.057792 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: E1205 12:07:14.069750 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.073470 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.073542 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.073558 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.073576 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.073591 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: E1205 12:07:14.085822 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:14 crc kubenswrapper[4807]: E1205 12:07:14.085938 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.087838 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.087883 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.087896 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.087915 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.087928 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.190390 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.190446 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.190454 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.190468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.190479 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.235309 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:14 crc kubenswrapper[4807]: E1205 12:07:14.235517 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.293591 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.293637 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.293652 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.293668 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.293680 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.396201 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.396242 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.396257 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.396276 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.396287 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.500373 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.500434 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.500450 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.500468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.500515 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.603505 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.603577 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.603588 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.603606 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.603617 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.705686 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.705734 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.705746 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.705765 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.705777 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.807896 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.807951 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.807968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.807988 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.808001 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.910670 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.910717 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.910730 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.910748 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:14 crc kubenswrapper[4807]: I1205 12:07:14.910761 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:14Z","lastTransitionTime":"2025-12-05T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.013726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.013798 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.013817 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.013841 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.013860 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.116533 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.116575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.116589 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.116610 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.116630 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.219567 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.219621 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.219635 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.219652 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.219663 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.234998 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.234999 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.235157 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:15 crc kubenswrapper[4807]: E1205 12:07:15.235288 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:15 crc kubenswrapper[4807]: E1205 12:07:15.235419 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:15 crc kubenswrapper[4807]: E1205 12:07:15.235482 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.322591 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.322648 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.322669 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.322694 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.322711 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.425346 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.425414 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.425428 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.425447 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.425459 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.528231 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.528285 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.528299 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.528317 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.528327 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.630505 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.630615 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.630626 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.630641 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.630653 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.733319 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.733382 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.733392 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.733424 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.733436 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.835616 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.835654 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.835670 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.835685 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.835695 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.937985 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.938049 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.938065 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.938083 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:15 crc kubenswrapper[4807]: I1205 12:07:15.938094 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:15Z","lastTransitionTime":"2025-12-05T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.041058 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.041130 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.041148 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.041174 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.041195 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.144101 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.144585 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.144601 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.144620 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.144657 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.235292 4807 scope.go:117] "RemoveContainer" containerID="88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f" Dec 05 12:07:16 crc kubenswrapper[4807]: E1205 12:07:16.235458 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.235657 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:16 crc kubenswrapper[4807]: E1205 12:07:16.235747 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.247086 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.247133 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.247147 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.247163 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.247174 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.349844 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.349884 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.349894 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.349911 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.349922 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.452021 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.452064 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.452082 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.452099 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.452109 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.554649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.554704 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.554717 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.554738 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.554754 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.656313 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.656384 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.656417 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.656427 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.656442 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: E1205 12:07:16.656456 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.656453 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: E1205 12:07:16.656542 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:07:48.65650511 +0000 UTC m=+98.150368379 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.758953 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.759006 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.759015 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.759028 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.759037 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.860940 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.860972 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.861011 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.861027 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.861041 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.963484 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.963541 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.963554 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.963571 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:16 crc kubenswrapper[4807]: I1205 12:07:16.963581 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:16Z","lastTransitionTime":"2025-12-05T12:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.065688 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.065725 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.065738 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.065755 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.065767 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.168926 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.168973 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.168990 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.169015 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.169031 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.235050 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.235094 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.235192 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:17 crc kubenswrapper[4807]: E1205 12:07:17.235370 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:17 crc kubenswrapper[4807]: E1205 12:07:17.235454 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:17 crc kubenswrapper[4807]: E1205 12:07:17.235667 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.274286 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.274360 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.274374 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.274400 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.274424 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.377346 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.377406 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.377423 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.377448 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.377477 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.479412 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.479453 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.479464 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.479478 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.479489 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.581849 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.581891 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.581903 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.581919 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.581931 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.684626 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.684657 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.684667 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.684682 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.684693 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.787170 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.787219 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.787229 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.787244 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.787254 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.889988 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.890060 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.890072 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.890089 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.890104 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.992751 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.992803 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.992815 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.992832 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:17 crc kubenswrapper[4807]: I1205 12:07:17.992844 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:17Z","lastTransitionTime":"2025-12-05T12:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.095598 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.095663 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.095674 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.095692 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.095706 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.198586 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.198649 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.198663 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.198683 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.198695 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.235473 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:18 crc kubenswrapper[4807]: E1205 12:07:18.235708 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.301303 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.301370 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.301385 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.301410 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.301430 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.403732 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.403785 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.403795 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.403811 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.403827 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.507241 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.507290 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.507302 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.507321 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.507334 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.609200 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.609232 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.609241 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.609258 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.609267 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.655498 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/0.log" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.655584 4807 generic.go:334] "Generic (PLEG): container finished" podID="379e7e5d-522d-409b-84db-fafbd1bff182" containerID="3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8" exitCode=1 Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.655630 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerDied","Data":"3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.656091 4807 scope.go:117] "RemoveContainer" containerID="3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.682756 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.693349 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.711965 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.712003 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.712013 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.712082 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.712095 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.713227 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.723754 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24431c1-0ace-441e-be42-4b199d46972b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.735705 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.746019 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.758811 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.770751 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"2025-12-05T12:06:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2\\\\n2025-12-05T12:06:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2 to /host/opt/cni/bin/\\\\n2025-12-05T12:06:32Z [verbose] multus-daemon started\\\\n2025-12-05T12:06:32Z [verbose] Readiness Indicator file check\\\\n2025-12-05T12:07:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.783045 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.798304 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.814849 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.816063 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.816104 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.816115 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.816167 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.816180 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.828148 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.838186 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.850477 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.861734 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.872131 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.882721 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.895133 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:18Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.918250 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.918295 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.918306 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.918323 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:18 crc kubenswrapper[4807]: I1205 12:07:18.918338 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:18Z","lastTransitionTime":"2025-12-05T12:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.020797 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.020835 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.020847 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.020863 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.020874 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.123084 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.123126 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.123135 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.123150 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.123161 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.225014 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.225056 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.225068 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.225086 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.225099 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.234609 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.234654 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.234695 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:19 crc kubenswrapper[4807]: E1205 12:07:19.234825 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:19 crc kubenswrapper[4807]: E1205 12:07:19.234930 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:19 crc kubenswrapper[4807]: E1205 12:07:19.235092 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.327261 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.327311 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.327323 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.327339 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.327353 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.429841 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.429884 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.429894 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.429909 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.429918 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.532426 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.532468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.532479 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.532497 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.532507 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.636007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.636059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.636068 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.636084 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.636094 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.661478 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/0.log" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.661578 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerStarted","Data":"ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.683795 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.701910 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.715103 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.727395 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"2025-12-05T12:06:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2\\\\n2025-12-05T12:06:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2 to /host/opt/cni/bin/\\\\n2025-12-05T12:06:32Z [verbose] multus-daemon started\\\\n2025-12-05T12:06:32Z [verbose] Readiness Indicator file check\\\\n2025-12-05T12:07:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.736640 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24431c1-0ace-441e-be42-4b199d46972b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.738161 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.738183 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.738191 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.738205 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.738213 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.749602 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.762193 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.773709 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.790017 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.801315 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.811046 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.825244 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.834952 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.839977 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.840003 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.840012 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.840025 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.840035 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.847295 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.859255 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.878437 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.894549 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.906158 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.942001 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.942043 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.942053 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.942090 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:19 crc kubenswrapper[4807]: I1205 12:07:19.942103 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:19Z","lastTransitionTime":"2025-12-05T12:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.045129 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.045169 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.045181 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.045194 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.045205 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.148007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.148237 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.148331 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.148396 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.148463 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.234422 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:20 crc kubenswrapper[4807]: E1205 12:07:20.234587 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.251644 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.251745 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.251776 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.251803 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.251820 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.355050 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.355132 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.355156 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.355183 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.355205 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.457538 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.457578 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.457589 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.457607 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.457616 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.560736 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.560786 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.560798 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.560816 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.560827 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.663728 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.663770 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.663781 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.663798 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.663811 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.766508 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.766567 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.766579 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.766598 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.766610 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.869041 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.869099 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.869117 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.869142 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.869159 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.970878 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.970916 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.970926 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.970942 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:20 crc kubenswrapper[4807]: I1205 12:07:20.970954 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:20Z","lastTransitionTime":"2025-12-05T12:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.073156 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.073212 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.073221 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.073235 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.073246 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.175957 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.175995 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.176003 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.176019 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.176029 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.235021 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.235106 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:21 crc kubenswrapper[4807]: E1205 12:07:21.235184 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.235126 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:21 crc kubenswrapper[4807]: E1205 12:07:21.235355 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:21 crc kubenswrapper[4807]: E1205 12:07:21.235442 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.248251 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.258313 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.271214 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.278156 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.278196 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.278204 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.278218 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.278228 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.282885 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.294442 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.306265 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.324176 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.344126 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.356718 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24431c1-0ace-441e-be42-4b199d46972b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.371932 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.380411 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.380445 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.380455 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.380472 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.380483 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.385865 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.399105 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.411772 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"2025-12-05T12:06:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2\\\\n2025-12-05T12:06:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2 to /host/opt/cni/bin/\\\\n2025-12-05T12:06:32Z [verbose] multus-daemon started\\\\n2025-12-05T12:06:32Z [verbose] Readiness Indicator file check\\\\n2025-12-05T12:07:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.428057 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.442518 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.455518 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.465543 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.478772 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.482314 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.482357 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.482370 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.482387 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.482399 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.584242 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.584276 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.584285 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.584301 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.584311 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.686739 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.686821 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.686839 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.686867 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.686889 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.789215 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.789251 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.789264 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.789281 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.789295 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.892420 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.892487 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.892500 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.892551 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.892566 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.994913 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.994991 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.995001 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.995016 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:21 crc kubenswrapper[4807]: I1205 12:07:21.995027 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:21Z","lastTransitionTime":"2025-12-05T12:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.097075 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.097109 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.097120 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.097134 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.097145 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.199870 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.199929 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.199944 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.199963 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.199975 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.234644 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:22 crc kubenswrapper[4807]: E1205 12:07:22.234823 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.302887 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.302927 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.302937 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.302951 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.302962 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.405541 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.405591 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.405600 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.405615 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.405625 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.508210 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.508270 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.508284 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.508304 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.508318 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.611435 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.611501 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.611518 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.611569 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.611602 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.714468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.714511 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.714554 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.714569 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.714583 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.817814 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.817873 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.817886 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.817905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.817917 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.920646 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.920689 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.920698 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.920716 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:22 crc kubenswrapper[4807]: I1205 12:07:22.920727 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:22Z","lastTransitionTime":"2025-12-05T12:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.024445 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.024547 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.024575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.024607 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.024628 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.127210 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.127272 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.127289 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.127312 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.127330 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.229413 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.229459 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.229472 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.229489 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.229501 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.234762 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:23 crc kubenswrapper[4807]: E1205 12:07:23.234892 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.234980 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.234757 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:23 crc kubenswrapper[4807]: E1205 12:07:23.235191 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:23 crc kubenswrapper[4807]: E1205 12:07:23.235297 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.331417 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.331452 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.331460 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.331473 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.331482 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.434338 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.434388 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.434403 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.434420 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.434431 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.536568 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.536602 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.536617 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.536639 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.536650 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.639514 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.639585 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.639633 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.639652 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.639665 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.744964 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.745024 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.745038 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.745058 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.745076 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.847945 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.847994 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.848007 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.848026 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.848039 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.949996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.950059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.950077 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.950101 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:23 crc kubenswrapper[4807]: I1205 12:07:23.950118 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:23Z","lastTransitionTime":"2025-12-05T12:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.054372 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.054416 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.054437 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.054457 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.054474 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.157722 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.157771 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.157780 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.157796 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.157806 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.235131 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:24 crc kubenswrapper[4807]: E1205 12:07:24.235417 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.247736 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.248503 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.248577 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.248589 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.248606 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.248619 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: E1205 12:07:24.263118 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.267989 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.268025 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.268065 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.268083 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.268095 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: E1205 12:07:24.281838 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.284860 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.284895 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.284908 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.284923 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.284935 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: E1205 12:07:24.296712 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.299958 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.299987 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.299998 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.300012 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.300023 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: E1205 12:07:24.315947 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.319134 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.319181 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.319198 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.319220 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.319236 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: E1205 12:07:24.331053 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:24 crc kubenswrapper[4807]: E1205 12:07:24.331201 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.332963 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.333002 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.333012 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.333026 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.333037 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.434883 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.434912 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.434921 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.434934 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.434943 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.537433 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.537469 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.537478 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.537492 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.537501 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.640653 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.640714 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.640727 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.640745 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.640759 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.743233 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.743324 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.743341 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.743363 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.743378 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.845400 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.845458 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.845472 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.845492 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.845509 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.947930 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.947970 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.947980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.947996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:24 crc kubenswrapper[4807]: I1205 12:07:24.948008 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:24Z","lastTransitionTime":"2025-12-05T12:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.050433 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.050485 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.050498 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.050517 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.050544 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.153344 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.153391 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.153403 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.153423 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.153433 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.234703 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.234711 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:25 crc kubenswrapper[4807]: E1205 12:07:25.234880 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.234725 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:25 crc kubenswrapper[4807]: E1205 12:07:25.235082 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:25 crc kubenswrapper[4807]: E1205 12:07:25.235145 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.256745 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.256803 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.256816 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.256835 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.256848 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.359951 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.360026 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.360047 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.360077 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.360094 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.463413 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.463456 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.463468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.463486 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.463499 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.565996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.566037 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.566059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.566081 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.566096 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.668770 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.668828 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.668843 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.668862 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.668875 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.771577 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.771620 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.771629 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.771643 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.771653 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.874182 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.874278 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.874297 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.874321 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.874338 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.977484 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.977571 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.977591 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.977623 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:25 crc kubenswrapper[4807]: I1205 12:07:25.977659 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:25Z","lastTransitionTime":"2025-12-05T12:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.079951 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.080052 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.080069 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.080093 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.080110 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.183708 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.183748 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.183757 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.183774 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.183786 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.234655 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:26 crc kubenswrapper[4807]: E1205 12:07:26.234862 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.285883 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.285936 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.285948 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.285968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.285983 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.388155 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.388216 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.388227 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.388247 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.388262 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.491199 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.491282 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.491307 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.491336 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.491357 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.595155 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.595196 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.595207 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.595222 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.595235 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.696850 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.696900 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.696911 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.696926 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.696938 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.798631 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.798673 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.798684 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.798696 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.798706 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.900609 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.900675 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.900709 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.900730 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:26 crc kubenswrapper[4807]: I1205 12:07:26.900744 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:26Z","lastTransitionTime":"2025-12-05T12:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.003819 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.003945 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.003965 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.003991 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.004008 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.106322 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.106368 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.106378 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.106396 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.106408 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.208650 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.208695 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.208709 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.208723 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.208734 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.234980 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.235035 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:27 crc kubenswrapper[4807]: E1205 12:07:27.235142 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.234980 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:27 crc kubenswrapper[4807]: E1205 12:07:27.235234 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:27 crc kubenswrapper[4807]: E1205 12:07:27.235284 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.311235 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.311296 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.311315 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.311337 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.311352 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.414662 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.414732 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.414749 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.414774 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.414791 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.518307 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.518384 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.518411 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.518440 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.518460 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.621840 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.621915 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.621937 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.621964 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.621986 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.724998 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.725059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.725075 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.725097 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.725109 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.828210 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.828249 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.828258 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.828271 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.828280 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.930795 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.930863 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.930881 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.930904 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:27 crc kubenswrapper[4807]: I1205 12:07:27.930921 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:27Z","lastTransitionTime":"2025-12-05T12:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.033004 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.033049 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.033059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.033075 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.033089 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.135321 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.135396 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.135418 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.135459 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.135485 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.234755 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:28 crc kubenswrapper[4807]: E1205 12:07:28.235016 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.238265 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.238318 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.238330 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.238349 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.238364 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.341149 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.341202 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.341212 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.341229 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.341239 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.443890 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.443952 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.443967 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.444017 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.444038 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.546768 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.546815 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.546823 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.546840 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.546850 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.650068 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.650119 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.650129 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.650152 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.650165 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.753186 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.753263 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.753478 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.753509 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.753566 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.856716 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.856788 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.856803 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.856822 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.856834 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.959215 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.959280 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.959297 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.959322 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:28 crc kubenswrapper[4807]: I1205 12:07:28.959337 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:28Z","lastTransitionTime":"2025-12-05T12:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.062199 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.062298 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.062318 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.062348 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.062371 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.164443 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.164499 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.164514 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.164563 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.164585 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.237404 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.237450 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.237828 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:29 crc kubenswrapper[4807]: E1205 12:07:29.237921 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:29 crc kubenswrapper[4807]: E1205 12:07:29.238024 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:29 crc kubenswrapper[4807]: E1205 12:07:29.238149 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.238292 4807 scope.go:117] "RemoveContainer" containerID="88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.266982 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.267295 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.267305 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.267319 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.267330 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.369176 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.369214 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.369222 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.369235 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.369244 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.471420 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.471457 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.471466 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.471479 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.471489 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.574144 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.574210 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.574223 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.574241 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.574254 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.676548 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.676597 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.676608 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.676628 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.676641 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.779642 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.779708 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.779726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.779759 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.779776 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.883362 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.883439 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.883468 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.883499 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.883570 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.986361 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.986428 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.986440 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.986455 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:29 crc kubenswrapper[4807]: I1205 12:07:29.986465 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:29Z","lastTransitionTime":"2025-12-05T12:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.089953 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.090003 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.090012 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.090027 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.090038 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.192178 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.192213 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.192222 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.192235 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.192244 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.235359 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:30 crc kubenswrapper[4807]: E1205 12:07:30.235563 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.294777 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.295033 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.295047 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.295067 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.295079 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.397419 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.397475 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.397486 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.397501 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.397511 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.500594 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.500636 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.500645 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.500660 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.500670 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.602425 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.602467 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.602476 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.602491 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.602499 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.699334 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/2.log" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.703159 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.704437 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.704470 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.704479 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.704492 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.704502 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.809157 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.809199 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.809209 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.809223 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.809233 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.911943 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.912009 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.912022 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.912041 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:30 crc kubenswrapper[4807]: I1205 12:07:30.912055 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:30Z","lastTransitionTime":"2025-12-05T12:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.015844 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.015883 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.015894 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.015910 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.015921 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.118198 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.118245 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.118258 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.118273 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.118286 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.220925 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.220964 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.220974 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.220988 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.220997 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.235314 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.235323 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:31 crc kubenswrapper[4807]: E1205 12:07:31.235463 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:31 crc kubenswrapper[4807]: E1205 12:07:31.235576 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.235340 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:31 crc kubenswrapper[4807]: E1205 12:07:31.235672 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.256383 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.268099 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.288028 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.302486 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.314208 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"2025-12-05T12:06:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2\\\\n2025-12-05T12:06:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2 to /host/opt/cni/bin/\\\\n2025-12-05T12:06:32Z [verbose] multus-daemon started\\\\n2025-12-05T12:06:32Z [verbose] Readiness Indicator file check\\\\n2025-12-05T12:07:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.323583 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.323637 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.323651 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.323667 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.323679 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.325082 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b92b6ebb-ae93-4101-b9e4-6bb5f02523c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726759440c5213ec731623e92b8fce4760be11e904289c1dae4a5103b5af62e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.336875 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24431c1-0ace-441e-be42-4b199d46972b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.348491 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.360683 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.372694 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.384812 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.395469 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.406150 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.416484 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.425654 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.425716 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.425728 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.425750 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.425760 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.426583 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.436788 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.445272 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.453975 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.462480 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.528838 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.528889 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.528904 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.528920 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.528928 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.630897 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.630951 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.630962 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.630979 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.630992 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.709966 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/3.log" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.710674 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/2.log" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.713552 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" exitCode=1 Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.713603 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.713659 4807 scope.go:117] "RemoveContainer" containerID="88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.714518 4807 scope.go:117] "RemoveContainer" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" Dec 05 12:07:31 crc kubenswrapper[4807]: E1205 12:07:31.714754 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.734179 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.734230 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.734243 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.734260 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.734273 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.736791 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b92b6ebb-ae93-4101-b9e4-6bb5f02523c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726759440c5213ec731623e92b8fce4760be11e904289c1dae4a5103b5af62e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.796875 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24431c1-0ace-441e-be42-4b199d46972b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.812107 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.826383 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.835900 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.835938 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.835952 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.835970 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.835986 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.844316 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.856511 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"2025-12-05T12:06:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2\\\\n2025-12-05T12:06:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2 to /host/opt/cni/bin/\\\\n2025-12-05T12:06:32Z [verbose] multus-daemon started\\\\n2025-12-05T12:06:32Z [verbose] Readiness Indicator file check\\\\n2025-12-05T12:07:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.870588 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.882762 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.895830 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.908752 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.920083 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.932785 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.937910 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.937955 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.937970 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.937989 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.938005 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:31Z","lastTransitionTime":"2025-12-05T12:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.942788 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.952895 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.963224 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:31 crc kubenswrapper[4807]: I1205 12:07:31.975392 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:31.999884 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.012016 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.032071 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88840449f6ecc78b8059cb2a58e706c566283db4229d16612df242c00df29b5f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:00Z\\\",\\\"message\\\":\\\" service openshift-dns/dns-default for network=default\\\\nI1205 12:07:00.102424 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102436 6474 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102260 6474 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-kth9r\\\\nI1205 12:07:00.102447 6474 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1205 12:07:00.102456 6474 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1205 12:07:00.102463 6474 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1205 12:07:00.102430 6474 services_controller.go:451] Built service openshift-network-console/networking-console-plugin cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Op\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:31Z\\\",\\\"message\\\":\\\".217.5.214:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {20da2226-531c-4179-9810-aa4026995ca3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:07:30.943242 6858 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:07:30.943263 6858 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-daemon cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-daemon_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}, Opts:services.LBOpts\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.039558 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.039588 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.039597 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.039612 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.039625 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.142659 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.142711 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.142722 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.142739 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.142750 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.235461 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:32 crc kubenswrapper[4807]: E1205 12:07:32.235730 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.245653 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.245740 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.245767 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.245796 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.245821 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.348871 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.348926 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.348948 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.348968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.348981 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.451972 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.452016 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.452028 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.452045 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.452056 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.554810 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.554862 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.554873 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.554887 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.554896 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.657195 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.657878 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.657980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.658083 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.658189 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.722863 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/3.log" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.726337 4807 scope.go:117] "RemoveContainer" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" Dec 05 12:07:32 crc kubenswrapper[4807]: E1205 12:07:32.726498 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.744353 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c96d5029ec1610b9b3ad3457e7cc1d87d20b79dd2ed819dc8644a83ad308b950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e006099db0953f10d95a1f8b35cd75649fb5afd379f31a42df6c4c7a121604fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.754855 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6eaedc27-e308-43e8-b3a9-14d094f179c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a0ec60d63a1fc653da74a29eaf151e0c558c3349969b4a113d5f8566de15022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da210d6f4b962be1b9662dbb3accfe4a1d706c0bcd5742c74be1acdfa61903a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdrrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2n72f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.760938 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.760970 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.760978 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.760993 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.761002 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.767262 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a27a2e4-e6d4-4de5-b18a-a7e806069a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 12:06:24.623873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-265226994/tls.crt::/tmp/serving-cert-265226994/tls.key\\\\\\\"\\\\nI1205 12:06:29.916600 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 12:06:29.918831 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 12:06:29.918852 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 12:06:29.918874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 12:06:29.918880 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 12:06:29.927042 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 12:06:29.927084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927090 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 12:06:29.927095 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 12:06:29.927100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 12:06:29.927105 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 12:06:29.927108 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 12:06:29.927428 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1205 12:06:29.936281 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1205 12:06:29.936352 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nF1205 12:06:29.936366 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.778864 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ee19a8d-3fcc-4760-bbd7-c255b0f8713e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e154932db454d1fc0305e715b0eb136f09a0b0c945a214b76a81a1cd53c6881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aad7b0b61640eb77bb63cff3073469e7644c90e08beb295ee8e00a1ffe76349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58486e3dea3b30aa4b783143d76b4c24f5ef092ed97d704aa44e82e1b9e0dcd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.791456 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.801128 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a779882f-2b7e-4ae0-addd-686fd4343bb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef938bfda3ca03579bb627efa59286d5a71ecc9f3c72f4f56958f5b414a41f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwdrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kth9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.811885 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb825b-4e77-44f1-a14b-5a3660dd1799\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ll4br\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4t4l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.825262 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.836466 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bnj26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e5318b9-e4e5-42ff-acef-811e7440033b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fe766f7aa7cfb3d5b7f351b204483c496f7a9b99ffb629e72a2c4832b2f4268\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bnj26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.847876 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wd694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1360d4c-bc7f-4557-9267-5fcfc052d9ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74ad24968025abbe5f41f2e092d407a4ffc51343e9e685e7794bbabbbb9a30fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hl8hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wd694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.864609 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.864713 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.864779 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.864804 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.864860 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.868613 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fac0aac-e8ed-4d24-883f-1d2e756ff36a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://228b2609fc92d9dbd6a917be5c43595e935f4285aa4387dd7bbf715ce37cddd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a996118707d9726e8195193fbaabfc05dec491ba07f430e820a72e7838e7ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6c81ee28e3261dd6588dbdba9f33f7cd98018b91098fee2c9809fda407e706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c4f1c765b709a604eafa569091cd326f2ecb9ec1483b8682c6a3155329948a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61c54a27f5832340918222102c8e03c0d2e6c2be2294a9c92823219520a26a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5d5cc5807074078de998077ee3dfe022acc9506fd333ab07244e79747e2c44a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711ef02e94d9f8885123939db7246ac9c6015c7d0bb1a215903a4028d87be0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fe865ae68c943f7174344dc0c8de275dae60a97b8214510274efb0ccafe2f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.879296 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231f4d76ade495e5674e76f7a6ae23ab54a1850ebbd477ee8bda944b5f99dd0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.911309 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9bb753e-da01-4d90-95b2-5614ab510ec8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:31Z\\\",\\\"message\\\":\\\".217.5.214:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {20da2226-531c-4179-9810-aa4026995ca3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:07:30.943242 6858 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 12:07:30.943263 6858 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-daemon cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-daemon_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}, Opts:services.LBOpts\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:07:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g748d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-84bd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.922372 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.922801 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.937553 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f23809b6-3d63-484f-8e7c-9dc5dac87ebd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ce33501207553c21c677e4c858184f19300b64fb23913556806df5c48fd6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67fbdb92ad16a772f63ad27a0a189e89f0d51ccf9a0b926d53f2ef20063376\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a0b56c5f6a65c72eee7e802d0f877abb6d9016a9f85fac77b90edbab8de4d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e0213aa742073df98287b68342df94f7bf5423757ee671cdd755dc95eabb61d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ef05f02c03e3ed79dd60b3d0713774ccfc76e247e2840ef52125ef938cd4a4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1832a1747e12db9384a9e30486a3c7de1fbb22b760dff6e5ad91a66009b506d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a164c96018b5cf6a827b625ad6969070d0a096e61be1bd9b27aed48406d15db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5j59f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tsqzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.951118 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sq7cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"379e7e5d-522d-409b-84db-fafbd1bff182\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T12:07:18Z\\\",\\\"message\\\":\\\"2025-12-05T12:06:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2\\\\n2025-12-05T12:06:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_822204c3-0b50-4fa8-9408-10b988edb0f2 to /host/opt/cni/bin/\\\\n2025-12-05T12:06:32Z [verbose] multus-daemon started\\\\n2025-12-05T12:06:32Z [verbose] Readiness Indicator file check\\\\n2025-12-05T12:07:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-svt9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sq7cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.959914 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b92b6ebb-ae93-4101-b9e4-6bb5f02523c2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6726759440c5213ec731623e92b8fce4760be11e904289c1dae4a5103b5af62e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbcd552ec90e7d6d9dbc2e11d40d079acbeb3d8500dbb4b8cf87cc9bec7b72ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.966945 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.966983 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.966996 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.967012 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.967023 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:32Z","lastTransitionTime":"2025-12-05T12:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.969205 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24431c1-0ace-441e-be42-4b199d46972b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2459f97951a26b59add58d37c11fc7beea0c2ff0c8698ab9e579e7fc1ebdaecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0d466a4c110fe3ab3f1bd2ee4b425b294c8df2b2089d8fe8698421a59dfb522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc1eb7aa374d31efd4aa4a966a7135a5ac8013b0c031dc0bbaaf9bfe174a702d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15da7eff7c3ee89e8f9e70142ddc4cd8c498d10a8d5a8f1e8b0d815a08db777f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T12:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T12:06:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T12:06:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:32 crc kubenswrapper[4807]: I1205 12:07:32.981202 4807 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T12:06:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84356085b3a8eb5d2c62029e88f71fcee19c446ee25dc3ce1a1596615fd21135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T12:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:32Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.070210 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.070271 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.070296 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.070318 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.070331 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.173475 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.173578 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.173606 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.173633 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.173654 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.234705 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.234810 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.234908 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.234961 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.235170 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.235316 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.277013 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.277061 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.277074 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.277092 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.277103 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.380496 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.380767 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.380777 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.380793 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.380803 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.483484 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.483551 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.483562 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.483575 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.483586 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.585925 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.585981 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.585991 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.586008 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.586018 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.688447 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.688495 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.688505 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.688537 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.688549 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.732027 4807 scope.go:117] "RemoveContainer" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.732264 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.740106 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.740283 4807 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.740389 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.740367726 +0000 UTC m=+147.234231095 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.791652 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.791701 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.791712 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.791727 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.791737 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.841579 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.841705 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.841775 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.841748692 +0000 UTC m=+147.335611961 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.841882 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.841910 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.841932 4807 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.841952 4807 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.841995 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.841973667 +0000 UTC m=+147.335836966 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.842038 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.842025678 +0000 UTC m=+147.335888987 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.841909 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.842116 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.842221 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.842240 4807 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.842254 4807 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:33 crc kubenswrapper[4807]: E1205 12:07:33.842342 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.842334066 +0000 UTC m=+147.336197325 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.894846 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.894878 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.894897 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.894915 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.894927 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.998595 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.998658 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.998675 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.998702 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:33 crc kubenswrapper[4807]: I1205 12:07:33.998727 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:33Z","lastTransitionTime":"2025-12-05T12:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.101905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.102262 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.102487 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.102805 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.103048 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.205509 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.205737 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.205817 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.205925 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.206018 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.234872 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:34 crc kubenswrapper[4807]: E1205 12:07:34.235021 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.308928 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.308968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.308980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.308995 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.309005 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.344090 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.344361 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.344370 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.344383 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.344391 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: E1205 12:07:34.357589 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.361731 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.361765 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.361774 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.361788 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.361797 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: E1205 12:07:34.376863 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.380913 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.380969 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.380981 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.381006 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.381023 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: E1205 12:07:34.396199 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.402024 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.402061 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.402070 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.402087 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.402096 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: E1205 12:07:34.419599 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.424294 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.424342 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.424356 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.424383 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.424396 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: E1205 12:07:34.438213 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T12:07:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b68900d3-0b24-4276-b2e2-04428fc986a2\\\",\\\"systemUUID\\\":\\\"46e7cf4e-5f92-4473-b811-c45109d1ff45\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T12:07:34Z is after 2025-08-24T17:21:41Z" Dec 05 12:07:34 crc kubenswrapper[4807]: E1205 12:07:34.438367 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.439802 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.439841 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.439854 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.439869 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.439879 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.542853 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.543134 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.543299 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.543460 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.543624 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.646619 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.646702 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.646720 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.646744 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.646760 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.748855 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.748905 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.748918 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.748935 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.748947 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.852510 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.852585 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.852598 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.852615 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.852630 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.955773 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.955843 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.955863 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.955889 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:34 crc kubenswrapper[4807]: I1205 12:07:34.955913 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:34Z","lastTransitionTime":"2025-12-05T12:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.058489 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.058596 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.058619 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.058646 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.058666 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.161949 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.162031 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.162043 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.162059 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.162071 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.234667 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.234708 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:35 crc kubenswrapper[4807]: E1205 12:07:35.234808 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.234840 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:35 crc kubenswrapper[4807]: E1205 12:07:35.234973 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:35 crc kubenswrapper[4807]: E1205 12:07:35.235077 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.265140 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.265193 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.265207 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.265227 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.265245 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.368271 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.368310 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.368318 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.368334 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.368345 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.471293 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.471341 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.471354 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.471373 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.471385 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.574050 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.574142 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.574167 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.574196 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.574217 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.676375 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.676427 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.676437 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.676454 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.676463 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.783102 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.783172 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.783190 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.783215 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.783235 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.885907 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.886177 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.886253 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.886343 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.886428 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.989333 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.989657 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.989829 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.989980 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:35 crc kubenswrapper[4807]: I1205 12:07:35.990139 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:35Z","lastTransitionTime":"2025-12-05T12:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.092581 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.092610 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.092618 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.092630 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.092639 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.194883 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.194916 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.194925 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.194938 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.194947 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.235184 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:36 crc kubenswrapper[4807]: E1205 12:07:36.235640 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.297051 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.297098 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.297110 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.297127 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.297139 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.400381 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.400416 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.400425 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.400438 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.400448 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.502984 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.503290 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.503364 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.503445 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.503513 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.605700 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.605772 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.605789 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.605814 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.605831 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.709178 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.709257 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.709279 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.709307 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.709327 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.812405 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.813312 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.813605 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.813818 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.814026 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.916041 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.916084 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.916095 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.916111 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:36 crc kubenswrapper[4807]: I1205 12:07:36.916124 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:36Z","lastTransitionTime":"2025-12-05T12:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.018646 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.018684 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.018697 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.018716 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.018730 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.121302 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.121364 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.121381 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.121403 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.121418 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.224159 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.224195 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.224205 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.224257 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.224274 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.234672 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.234738 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.234744 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:37 crc kubenswrapper[4807]: E1205 12:07:37.234923 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:37 crc kubenswrapper[4807]: E1205 12:07:37.235021 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:37 crc kubenswrapper[4807]: E1205 12:07:37.235184 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.327036 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.327085 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.327096 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.327113 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.327135 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.429843 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.429898 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.429911 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.429930 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.429943 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.531988 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.532047 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.532061 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.532077 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.532087 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.634968 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.634998 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.635006 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.635021 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.635029 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.737241 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.737283 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.737296 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.737313 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.737324 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.840166 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.840237 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.840252 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.840278 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.840307 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.943436 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.943484 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.943495 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.943512 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:37 crc kubenswrapper[4807]: I1205 12:07:37.943540 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:37Z","lastTransitionTime":"2025-12-05T12:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.045962 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.046009 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.046020 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.046036 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.046046 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.149344 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.149415 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.149432 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.149455 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.149470 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.235140 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:38 crc kubenswrapper[4807]: E1205 12:07:38.235304 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.258414 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.258605 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.258651 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.258677 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.258697 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.364165 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.364207 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.364219 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.364236 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.364248 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.467142 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.467185 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.467196 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.467211 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.467222 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.569904 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.569974 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.569991 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.570016 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.570035 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.672708 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.672783 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.672804 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.672833 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.672850 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.775497 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.775545 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.775554 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.775567 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.775576 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.878975 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.879010 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.879018 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.879033 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.879045 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.981318 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.981379 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.981398 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.981419 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:38 crc kubenswrapper[4807]: I1205 12:07:38.981436 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:38Z","lastTransitionTime":"2025-12-05T12:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.084774 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.084818 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.084830 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.084850 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.084862 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.188353 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.188411 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.188435 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.188464 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.188483 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.235369 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.235470 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.235622 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:39 crc kubenswrapper[4807]: E1205 12:07:39.235612 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:39 crc kubenswrapper[4807]: E1205 12:07:39.235751 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:39 crc kubenswrapper[4807]: E1205 12:07:39.235856 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.291325 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.291358 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.291366 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.291382 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.291404 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.394174 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.394237 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.394258 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.394285 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.394310 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.496676 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.496720 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.496737 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.496760 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.496779 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.599569 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.599613 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.599625 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.599641 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.599653 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.702000 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.702039 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.702049 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.702064 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.702075 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.805371 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.805419 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.805454 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.805492 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.805501 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.908121 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.908192 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.908222 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.908244 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:39 crc kubenswrapper[4807]: I1205 12:07:39.908258 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:39Z","lastTransitionTime":"2025-12-05T12:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.010666 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.010707 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.010778 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.010799 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.010810 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.114060 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.114135 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.114173 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.114205 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.114231 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.216501 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.216572 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.216584 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.216603 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.216617 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.235248 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:40 crc kubenswrapper[4807]: E1205 12:07:40.235350 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.319048 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.319093 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.319111 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.319139 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.319155 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.421610 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.421679 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.421702 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.421733 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.421756 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.524859 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.524892 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.524901 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.524917 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.524925 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.628134 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.628203 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.628214 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.628230 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.628241 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.731397 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.731459 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.731477 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.731501 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.731520 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.834302 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.834374 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.834391 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.834417 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.834438 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.941383 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.941431 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.941442 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.941461 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:40 crc kubenswrapper[4807]: I1205 12:07:40.941473 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:40Z","lastTransitionTime":"2025-12-05T12:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.044042 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.044308 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.044424 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.044518 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.044641 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.147418 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.147474 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.147483 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.147501 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.147512 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.234459 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:41 crc kubenswrapper[4807]: E1205 12:07:41.234815 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.234640 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:41 crc kubenswrapper[4807]: E1205 12:07:41.235047 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.234608 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:41 crc kubenswrapper[4807]: E1205 12:07:41.235220 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.251057 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.251098 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.251108 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.251125 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.251135 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.294910 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-wd694" podStartSLOduration=71.294889264 podStartE2EDuration="1m11.294889264s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.294830032 +0000 UTC m=+90.788693331" watchObservedRunningTime="2025-12-05 12:07:41.294889264 +0000 UTC m=+90.788752543" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.295043 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-bnj26" podStartSLOduration=72.295037907 podStartE2EDuration="1m12.295037907s" podCreationTimestamp="2025-12-05 12:06:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.28527533 +0000 UTC m=+90.779138599" watchObservedRunningTime="2025-12-05 12:07:41.295037907 +0000 UTC m=+90.788901176" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.334166 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=71.334146226 podStartE2EDuration="1m11.334146226s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.332951038 +0000 UTC m=+90.826814307" watchObservedRunningTime="2025-12-05 12:07:41.334146226 +0000 UTC m=+90.828009505" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.334469 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podStartSLOduration=71.334458404 podStartE2EDuration="1m11.334458404s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.307110908 +0000 UTC m=+90.800974177" watchObservedRunningTime="2025-12-05 12:07:41.334458404 +0000 UTC m=+90.828321683" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.352753 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.352787 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.352795 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.352807 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.352816 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.438990 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-tsqzx" podStartSLOduration=70.438969722 podStartE2EDuration="1m10.438969722s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.424592908 +0000 UTC m=+90.918456197" watchObservedRunningTime="2025-12-05 12:07:41.438969722 +0000 UTC m=+90.932832991" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.439445 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-sq7cs" podStartSLOduration=71.439435783 podStartE2EDuration="1m11.439435783s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.439112255 +0000 UTC m=+90.932975524" watchObservedRunningTime="2025-12-05 12:07:41.439435783 +0000 UTC m=+90.933299052" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.455507 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.455560 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.455570 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.455584 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.455594 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.455604 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.455584398 podStartE2EDuration="17.455584398s" podCreationTimestamp="2025-12-05 12:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.455548957 +0000 UTC m=+90.949412226" watchObservedRunningTime="2025-12-05 12:07:41.455584398 +0000 UTC m=+90.949447667" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.498507 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=34.498489435 podStartE2EDuration="34.498489435s" podCreationTimestamp="2025-12-05 12:07:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.497898741 +0000 UTC m=+90.991762020" watchObservedRunningTime="2025-12-05 12:07:41.498489435 +0000 UTC m=+90.992352704" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.549858 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2n72f" podStartSLOduration=70.549840418 podStartE2EDuration="1m10.549840418s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.549601593 +0000 UTC m=+91.043464862" watchObservedRunningTime="2025-12-05 12:07:41.549840418 +0000 UTC m=+91.043703687" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.557937 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.557985 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.558000 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.558023 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.558038 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.565125 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.565106203 podStartE2EDuration="1m11.565106203s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.564172652 +0000 UTC m=+91.058035941" watchObservedRunningTime="2025-12-05 12:07:41.565106203 +0000 UTC m=+91.058969472" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.594464 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.594448215 podStartE2EDuration="1m8.594448215s" podCreationTimestamp="2025-12-05 12:06:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:41.576821926 +0000 UTC m=+91.070685215" watchObservedRunningTime="2025-12-05 12:07:41.594448215 +0000 UTC m=+91.088311474" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.660048 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.660083 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.660092 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.660105 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.660115 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.762499 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.762555 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.762565 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.762577 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.762586 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.865901 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.866199 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.866291 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.866401 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.866489 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.968936 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.968985 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.968999 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.969019 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:41 crc kubenswrapper[4807]: I1205 12:07:41.969034 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:41Z","lastTransitionTime":"2025-12-05T12:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.072786 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.072856 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.072874 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.072899 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.072920 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.175070 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.175130 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.175147 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.175174 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.175194 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.234441 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:42 crc kubenswrapper[4807]: E1205 12:07:42.234649 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.278013 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.278067 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.278084 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.278103 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.278116 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.381916 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.382162 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.382293 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.382362 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.382419 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.487857 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.487899 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.487910 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.487927 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.487938 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.590102 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.590167 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.590184 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.590203 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.590216 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.693423 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.693484 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.693494 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.693517 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.693554 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.796956 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.797036 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.797050 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.797072 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.797091 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.900647 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.900697 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.900706 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.900720 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:42 crc kubenswrapper[4807]: I1205 12:07:42.900731 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:42Z","lastTransitionTime":"2025-12-05T12:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.003664 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.003724 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.003760 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.003794 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.003818 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.107105 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.107173 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.107199 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.107230 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.107252 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.209870 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.209911 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.209921 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.209936 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.209944 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.235767 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.235833 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:43 crc kubenswrapper[4807]: E1205 12:07:43.235886 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:43 crc kubenswrapper[4807]: E1205 12:07:43.235950 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.235986 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:43 crc kubenswrapper[4807]: E1205 12:07:43.236050 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.312512 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.312564 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.312572 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.312588 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.312598 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.415343 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.415386 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.415397 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.415413 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.415425 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.517768 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.517821 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.517833 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.517848 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.517857 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.619932 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.619973 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.619985 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.620002 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.620015 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.722884 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.723285 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.723467 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.723726 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.723915 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.826334 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.826591 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.826705 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.826798 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.826887 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.929569 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.929809 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.929876 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.929953 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:43 crc kubenswrapper[4807]: I1205 12:07:43.930023 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:43Z","lastTransitionTime":"2025-12-05T12:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.031744 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.031805 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.031819 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.031837 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.031852 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:44Z","lastTransitionTime":"2025-12-05T12:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.134915 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.134967 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.134984 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.135003 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.135015 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:44Z","lastTransitionTime":"2025-12-05T12:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.235047 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:44 crc kubenswrapper[4807]: E1205 12:07:44.235630 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.237286 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.237340 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.237352 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.237371 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.237383 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:44Z","lastTransitionTime":"2025-12-05T12:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.339589 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.339845 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.339910 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.339978 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.340053 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:44Z","lastTransitionTime":"2025-12-05T12:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.448249 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.448317 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.448331 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.448349 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.448687 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:44Z","lastTransitionTime":"2025-12-05T12:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.509185 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.509469 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.509580 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.509689 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.509778 4807 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T12:07:44Z","lastTransitionTime":"2025-12-05T12:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.571611 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw"] Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.572317 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.575031 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.575209 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.575656 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.575702 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.667598 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.667655 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.667680 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.667733 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.667794 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.768882 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.768978 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.769014 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.769049 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.769109 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.769438 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.769433 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.770679 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.782465 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.786521 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aabb1c98-06e9-45c0-a48b-4115eda8c8eb-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nbfrw\" (UID: \"aabb1c98-06e9-45c0-a48b-4115eda8c8eb\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: I1205 12:07:44.887701 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" Dec 05 12:07:44 crc kubenswrapper[4807]: W1205 12:07:44.906765 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaabb1c98_06e9_45c0_a48b_4115eda8c8eb.slice/crio-f0602c9f9f18b8382a099804184702dd64a0fca1c15281cb5551564a5c955684 WatchSource:0}: Error finding container f0602c9f9f18b8382a099804184702dd64a0fca1c15281cb5551564a5c955684: Status 404 returned error can't find the container with id f0602c9f9f18b8382a099804184702dd64a0fca1c15281cb5551564a5c955684 Dec 05 12:07:45 crc kubenswrapper[4807]: I1205 12:07:45.234851 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:45 crc kubenswrapper[4807]: I1205 12:07:45.234945 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:45 crc kubenswrapper[4807]: E1205 12:07:45.235040 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:45 crc kubenswrapper[4807]: I1205 12:07:45.235127 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:45 crc kubenswrapper[4807]: E1205 12:07:45.235232 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:45 crc kubenswrapper[4807]: E1205 12:07:45.235360 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:45 crc kubenswrapper[4807]: I1205 12:07:45.767064 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" event={"ID":"aabb1c98-06e9-45c0-a48b-4115eda8c8eb","Type":"ContainerStarted","Data":"0f5cb792fb1e883da84efa34671cf099ac9daf8b3127fd978d990149644d76ff"} Dec 05 12:07:45 crc kubenswrapper[4807]: I1205 12:07:45.767125 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" event={"ID":"aabb1c98-06e9-45c0-a48b-4115eda8c8eb","Type":"ContainerStarted","Data":"f0602c9f9f18b8382a099804184702dd64a0fca1c15281cb5551564a5c955684"} Dec 05 12:07:46 crc kubenswrapper[4807]: I1205 12:07:46.235354 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:46 crc kubenswrapper[4807]: E1205 12:07:46.235602 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:47 crc kubenswrapper[4807]: I1205 12:07:47.234405 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:47 crc kubenswrapper[4807]: I1205 12:07:47.234504 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:47 crc kubenswrapper[4807]: E1205 12:07:47.234643 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:47 crc kubenswrapper[4807]: I1205 12:07:47.234742 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:47 crc kubenswrapper[4807]: E1205 12:07:47.234828 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:47 crc kubenswrapper[4807]: E1205 12:07:47.234999 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:47 crc kubenswrapper[4807]: I1205 12:07:47.235732 4807 scope.go:117] "RemoveContainer" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" Dec 05 12:07:47 crc kubenswrapper[4807]: E1205 12:07:47.235870 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:07:48 crc kubenswrapper[4807]: I1205 12:07:48.234869 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:48 crc kubenswrapper[4807]: E1205 12:07:48.235192 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:48 crc kubenswrapper[4807]: I1205 12:07:48.706956 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:48 crc kubenswrapper[4807]: E1205 12:07:48.707235 4807 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:07:48 crc kubenswrapper[4807]: E1205 12:07:48.707383 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs podName:a0fb825b-4e77-44f1-a14b-5a3660dd1799 nodeName:}" failed. No retries permitted until 2025-12-05 12:08:52.70734552 +0000 UTC m=+162.201208949 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs") pod "network-metrics-daemon-d4t4l" (UID: "a0fb825b-4e77-44f1-a14b-5a3660dd1799") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 12:07:49 crc kubenswrapper[4807]: I1205 12:07:49.235454 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:49 crc kubenswrapper[4807]: I1205 12:07:49.235708 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:49 crc kubenswrapper[4807]: I1205 12:07:49.235713 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:49 crc kubenswrapper[4807]: E1205 12:07:49.235991 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:49 crc kubenswrapper[4807]: E1205 12:07:49.235867 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:49 crc kubenswrapper[4807]: E1205 12:07:49.236096 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:50 crc kubenswrapper[4807]: I1205 12:07:50.234680 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:50 crc kubenswrapper[4807]: E1205 12:07:50.234894 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:51 crc kubenswrapper[4807]: I1205 12:07:51.234481 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:51 crc kubenswrapper[4807]: I1205 12:07:51.234480 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:51 crc kubenswrapper[4807]: I1205 12:07:51.234521 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:51 crc kubenswrapper[4807]: E1205 12:07:51.236223 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:51 crc kubenswrapper[4807]: E1205 12:07:51.236303 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:51 crc kubenswrapper[4807]: E1205 12:07:51.236889 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:52 crc kubenswrapper[4807]: I1205 12:07:52.234420 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:52 crc kubenswrapper[4807]: E1205 12:07:52.234597 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:53 crc kubenswrapper[4807]: I1205 12:07:53.235349 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:53 crc kubenswrapper[4807]: I1205 12:07:53.235728 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:53 crc kubenswrapper[4807]: I1205 12:07:53.235762 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:53 crc kubenswrapper[4807]: E1205 12:07:53.236332 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:53 crc kubenswrapper[4807]: E1205 12:07:53.236218 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:53 crc kubenswrapper[4807]: E1205 12:07:53.236124 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:54 crc kubenswrapper[4807]: I1205 12:07:54.234750 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:54 crc kubenswrapper[4807]: E1205 12:07:54.234888 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:55 crc kubenswrapper[4807]: I1205 12:07:55.235185 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:55 crc kubenswrapper[4807]: I1205 12:07:55.235411 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:55 crc kubenswrapper[4807]: E1205 12:07:55.236116 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:55 crc kubenswrapper[4807]: E1205 12:07:55.235995 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:55 crc kubenswrapper[4807]: I1205 12:07:55.235454 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:55 crc kubenswrapper[4807]: E1205 12:07:55.236210 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:56 crc kubenswrapper[4807]: I1205 12:07:56.234465 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:56 crc kubenswrapper[4807]: E1205 12:07:56.234944 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:57 crc kubenswrapper[4807]: I1205 12:07:57.234736 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:57 crc kubenswrapper[4807]: I1205 12:07:57.234740 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:57 crc kubenswrapper[4807]: I1205 12:07:57.234841 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:57 crc kubenswrapper[4807]: E1205 12:07:57.235833 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:57 crc kubenswrapper[4807]: E1205 12:07:57.235609 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:57 crc kubenswrapper[4807]: E1205 12:07:57.235472 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:58 crc kubenswrapper[4807]: I1205 12:07:58.235383 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:07:58 crc kubenswrapper[4807]: E1205 12:07:58.236647 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:07:59 crc kubenswrapper[4807]: I1205 12:07:59.234598 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:07:59 crc kubenswrapper[4807]: I1205 12:07:59.234620 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:07:59 crc kubenswrapper[4807]: E1205 12:07:59.234828 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:07:59 crc kubenswrapper[4807]: I1205 12:07:59.235215 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:07:59 crc kubenswrapper[4807]: E1205 12:07:59.235372 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:07:59 crc kubenswrapper[4807]: E1205 12:07:59.235667 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:07:59 crc kubenswrapper[4807]: I1205 12:07:59.235904 4807 scope.go:117] "RemoveContainer" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" Dec 05 12:07:59 crc kubenswrapper[4807]: E1205 12:07:59.236156 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-84bd8_openshift-ovn-kubernetes(a9bb753e-da01-4d90-95b2-5614ab510ec8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" Dec 05 12:08:00 crc kubenswrapper[4807]: I1205 12:08:00.235232 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:00 crc kubenswrapper[4807]: E1205 12:08:00.235944 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:01 crc kubenswrapper[4807]: I1205 12:08:01.234430 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:01 crc kubenswrapper[4807]: E1205 12:08:01.235681 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:01 crc kubenswrapper[4807]: I1205 12:08:01.235711 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:01 crc kubenswrapper[4807]: I1205 12:08:01.235774 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:01 crc kubenswrapper[4807]: E1205 12:08:01.236070 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:01 crc kubenswrapper[4807]: E1205 12:08:01.236171 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:02 crc kubenswrapper[4807]: I1205 12:08:02.235010 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:02 crc kubenswrapper[4807]: E1205 12:08:02.235248 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:03 crc kubenswrapper[4807]: I1205 12:08:03.235040 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:03 crc kubenswrapper[4807]: I1205 12:08:03.235319 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:03 crc kubenswrapper[4807]: I1205 12:08:03.235368 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:03 crc kubenswrapper[4807]: E1205 12:08:03.235472 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:03 crc kubenswrapper[4807]: E1205 12:08:03.235611 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:03 crc kubenswrapper[4807]: E1205 12:08:03.235673 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:04 crc kubenswrapper[4807]: I1205 12:08:04.234809 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:04 crc kubenswrapper[4807]: E1205 12:08:04.235019 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.235501 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.235511 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.235647 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:05 crc kubenswrapper[4807]: E1205 12:08:05.236234 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:05 crc kubenswrapper[4807]: E1205 12:08:05.236765 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:05 crc kubenswrapper[4807]: E1205 12:08:05.237036 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.843394 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/1.log" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.843930 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/0.log" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.844063 4807 generic.go:334] "Generic (PLEG): container finished" podID="379e7e5d-522d-409b-84db-fafbd1bff182" containerID="ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f" exitCode=1 Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.844125 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerDied","Data":"ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f"} Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.844213 4807 scope.go:117] "RemoveContainer" containerID="3ad58058884c9b1ba66c34d50819911415801714d064028453c9f7ebae2903f8" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.844564 4807 scope.go:117] "RemoveContainer" containerID="ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f" Dec 05 12:08:05 crc kubenswrapper[4807]: E1205 12:08:05.844735 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-sq7cs_openshift-multus(379e7e5d-522d-409b-84db-fafbd1bff182)\"" pod="openshift-multus/multus-sq7cs" podUID="379e7e5d-522d-409b-84db-fafbd1bff182" Dec 05 12:08:05 crc kubenswrapper[4807]: I1205 12:08:05.861492 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbfrw" podStartSLOduration=94.861477239 podStartE2EDuration="1m34.861477239s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:07:45.787349697 +0000 UTC m=+95.281212966" watchObservedRunningTime="2025-12-05 12:08:05.861477239 +0000 UTC m=+115.355340508" Dec 05 12:08:06 crc kubenswrapper[4807]: I1205 12:08:06.234694 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:06 crc kubenswrapper[4807]: E1205 12:08:06.234824 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:06 crc kubenswrapper[4807]: I1205 12:08:06.850340 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/1.log" Dec 05 12:08:07 crc kubenswrapper[4807]: I1205 12:08:07.235058 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:07 crc kubenswrapper[4807]: I1205 12:08:07.235557 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:07 crc kubenswrapper[4807]: I1205 12:08:07.235613 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:07 crc kubenswrapper[4807]: E1205 12:08:07.235498 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:07 crc kubenswrapper[4807]: E1205 12:08:07.235812 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:07 crc kubenswrapper[4807]: E1205 12:08:07.236171 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:08 crc kubenswrapper[4807]: I1205 12:08:08.234391 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:08 crc kubenswrapper[4807]: E1205 12:08:08.234614 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:09 crc kubenswrapper[4807]: I1205 12:08:09.234593 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:09 crc kubenswrapper[4807]: I1205 12:08:09.234864 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:09 crc kubenswrapper[4807]: E1205 12:08:09.234960 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:09 crc kubenswrapper[4807]: I1205 12:08:09.235097 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:09 crc kubenswrapper[4807]: E1205 12:08:09.235267 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:09 crc kubenswrapper[4807]: E1205 12:08:09.235396 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:10 crc kubenswrapper[4807]: I1205 12:08:10.234988 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:10 crc kubenswrapper[4807]: E1205 12:08:10.235349 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:11 crc kubenswrapper[4807]: E1205 12:08:11.197560 4807 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.234760 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:11 crc kubenswrapper[4807]: E1205 12:08:11.236238 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.236252 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.236329 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:11 crc kubenswrapper[4807]: E1205 12:08:11.237197 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.237918 4807 scope.go:117] "RemoveContainer" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" Dec 05 12:08:11 crc kubenswrapper[4807]: E1205 12:08:11.237665 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:11 crc kubenswrapper[4807]: E1205 12:08:11.842005 4807 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.869497 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/3.log" Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.873148 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerStarted","Data":"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac"} Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.873729 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:08:11 crc kubenswrapper[4807]: I1205 12:08:11.912980 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podStartSLOduration=100.912951919 podStartE2EDuration="1m40.912951919s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:11.910964353 +0000 UTC m=+121.404827632" watchObservedRunningTime="2025-12-05 12:08:11.912951919 +0000 UTC m=+121.406815188" Dec 05 12:08:12 crc kubenswrapper[4807]: I1205 12:08:12.234861 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:12 crc kubenswrapper[4807]: E1205 12:08:12.235261 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:12 crc kubenswrapper[4807]: I1205 12:08:12.481487 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-d4t4l"] Dec 05 12:08:12 crc kubenswrapper[4807]: I1205 12:08:12.876485 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:12 crc kubenswrapper[4807]: E1205 12:08:12.876649 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:13 crc kubenswrapper[4807]: I1205 12:08:13.235502 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:13 crc kubenswrapper[4807]: I1205 12:08:13.235603 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:13 crc kubenswrapper[4807]: E1205 12:08:13.235655 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:13 crc kubenswrapper[4807]: I1205 12:08:13.235507 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:13 crc kubenswrapper[4807]: E1205 12:08:13.235720 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:13 crc kubenswrapper[4807]: E1205 12:08:13.236063 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:14 crc kubenswrapper[4807]: I1205 12:08:14.235052 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:14 crc kubenswrapper[4807]: E1205 12:08:14.235259 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:15 crc kubenswrapper[4807]: I1205 12:08:15.234858 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:15 crc kubenswrapper[4807]: I1205 12:08:15.235055 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:15 crc kubenswrapper[4807]: I1205 12:08:15.235209 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:15 crc kubenswrapper[4807]: E1205 12:08:15.235200 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:15 crc kubenswrapper[4807]: E1205 12:08:15.235349 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:15 crc kubenswrapper[4807]: E1205 12:08:15.235516 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:16 crc kubenswrapper[4807]: I1205 12:08:16.234699 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:16 crc kubenswrapper[4807]: I1205 12:08:16.235226 4807 scope.go:117] "RemoveContainer" containerID="ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f" Dec 05 12:08:16 crc kubenswrapper[4807]: E1205 12:08:16.235493 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:16 crc kubenswrapper[4807]: E1205 12:08:16.843342 4807 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:08:17 crc kubenswrapper[4807]: I1205 12:08:17.235505 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:17 crc kubenswrapper[4807]: E1205 12:08:17.235703 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:17 crc kubenswrapper[4807]: I1205 12:08:17.235947 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:17 crc kubenswrapper[4807]: E1205 12:08:17.236000 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:17 crc kubenswrapper[4807]: I1205 12:08:17.236200 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:17 crc kubenswrapper[4807]: E1205 12:08:17.236253 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:17 crc kubenswrapper[4807]: I1205 12:08:17.896007 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/1.log" Dec 05 12:08:17 crc kubenswrapper[4807]: I1205 12:08:17.896096 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerStarted","Data":"69b6cc6856985e969167fcd41b27cdb5f70f73283793841888c7c1db6caf3497"} Dec 05 12:08:18 crc kubenswrapper[4807]: I1205 12:08:18.235276 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:18 crc kubenswrapper[4807]: E1205 12:08:18.235502 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:19 crc kubenswrapper[4807]: I1205 12:08:19.235476 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:19 crc kubenswrapper[4807]: I1205 12:08:19.235619 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:19 crc kubenswrapper[4807]: E1205 12:08:19.235776 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:19 crc kubenswrapper[4807]: I1205 12:08:19.235809 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:19 crc kubenswrapper[4807]: E1205 12:08:19.235952 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:19 crc kubenswrapper[4807]: E1205 12:08:19.236075 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:20 crc kubenswrapper[4807]: I1205 12:08:20.235053 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:20 crc kubenswrapper[4807]: E1205 12:08:20.235209 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4t4l" podUID="a0fb825b-4e77-44f1-a14b-5a3660dd1799" Dec 05 12:08:21 crc kubenswrapper[4807]: I1205 12:08:21.234808 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:21 crc kubenswrapper[4807]: I1205 12:08:21.234801 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:21 crc kubenswrapper[4807]: I1205 12:08:21.234855 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:21 crc kubenswrapper[4807]: E1205 12:08:21.235747 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 12:08:21 crc kubenswrapper[4807]: E1205 12:08:21.235935 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 12:08:21 crc kubenswrapper[4807]: E1205 12:08:21.236154 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 12:08:22 crc kubenswrapper[4807]: I1205 12:08:22.234559 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:22 crc kubenswrapper[4807]: I1205 12:08:22.236946 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 12:08:22 crc kubenswrapper[4807]: I1205 12:08:22.237353 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 12:08:23 crc kubenswrapper[4807]: I1205 12:08:23.235129 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:23 crc kubenswrapper[4807]: I1205 12:08:23.235140 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:23 crc kubenswrapper[4807]: I1205 12:08:23.235129 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:23 crc kubenswrapper[4807]: I1205 12:08:23.237832 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 12:08:23 crc kubenswrapper[4807]: I1205 12:08:23.237837 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 12:08:23 crc kubenswrapper[4807]: I1205 12:08:23.238106 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 12:08:23 crc kubenswrapper[4807]: I1205 12:08:23.239837 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.207491 4807 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.248871 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h2v8s"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.249637 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xrqgd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.249831 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.251002 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.252046 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.252826 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.259253 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.259580 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.260016 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.260203 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.260305 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.260502 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.260704 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.260940 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.261034 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.261405 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.261703 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.261768 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.261918 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.262078 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.262102 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.262167 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.262691 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.266482 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.273615 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-g62qt"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.275368 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.275582 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.277227 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.278338 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.281746 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.282016 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.282151 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.282303 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.281751 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.282781 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.282966 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.283129 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.283274 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.283405 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.283717 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.283743 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.283952 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.284622 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.290696 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.291889 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.302811 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.303639 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.306153 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h2v8s"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.306945 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.307571 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.308016 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.308438 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.309500 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.309712 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.309797 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.309870 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.310270 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.310384 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.310476 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.310587 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.311941 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.313689 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.314290 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.314451 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hs4lf"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.315189 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.317172 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-25wjt"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.317997 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-25wjt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.319583 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bchkg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.320313 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.321378 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p67q6"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.321877 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.322991 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.327225 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.327897 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.328171 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.328515 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.329036 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.329135 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.329150 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.327901 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.329187 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.333681 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.333973 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-b52pb"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334269 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-njg49"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334411 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334492 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skdzd\" (UniqueName: \"kubernetes.io/projected/b0241b9d-883d-4e59-8f4c-38bf4f8efc91-kube-api-access-skdzd\") pod \"downloads-7954f5f757-25wjt\" (UID: \"b0241b9d-883d-4e59-8f4c-38bf4f8efc91\") " pod="openshift-console/downloads-7954f5f757-25wjt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334564 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-image-import-ca\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334593 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-etcd-client\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334620 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334667 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-node-pullsecrets\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334691 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-audit-dir\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334712 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334413 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334721 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-audit\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334842 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-etcd-client\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334875 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-serving-cert\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334927 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/27f3af31-4516-4987-9dbd-deb33f197008-audit-dir\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334946 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qxqx\" (UniqueName: \"kubernetes.io/projected/c07bf6c8-f1de-4d88-b692-31dec4fe9777-kube-api-access-8qxqx\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334980 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzjrd\" (UniqueName: \"kubernetes.io/projected/27f3af31-4516-4987-9dbd-deb33f197008-kube-api-access-jzjrd\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334999 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmwrn\" (UniqueName: \"kubernetes.io/projected/5496b467-3482-4ec0-85a0-d01c76d4b522-kube-api-access-vmwrn\") pod \"multus-admission-controller-857f4d67dd-hs4lf\" (UID: \"5496b467-3482-4ec0-85a0-d01c76d4b522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335025 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-etcd-serving-ca\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335041 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sf2g\" (UniqueName: \"kubernetes.io/projected/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-kube-api-access-4sf2g\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335108 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334493 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334570 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334614 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335107 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335400 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07bf6c8-f1de-4d88-b692-31dec4fe9777-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335424 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-config\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335443 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-encryption-config\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335463 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-serving-cert\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335481 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335496 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-encryption-config\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335514 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07bf6c8-f1de-4d88-b692-31dec4fe9777-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335559 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-audit-policies\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.335580 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5496b467-3482-4ec0-85a0-d01c76d4b522-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hs4lf\" (UID: \"5496b467-3482-4ec0-85a0-d01c76d4b522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334655 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.334694 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.336909 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337015 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337094 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337157 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337189 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337163 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337312 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337385 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337466 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.337717 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.338971 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.352767 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.353281 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.353447 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.354520 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.354666 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.356306 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.356432 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dnrjc"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.370709 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kf4mv"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.371402 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.372201 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.372470 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.372708 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.373847 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.376887 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.378053 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.379241 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.388516 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.388762 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.388926 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.389388 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.389580 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.389718 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.390212 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.390406 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.390620 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.390863 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.389602 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.392744 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.392985 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393056 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393224 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393268 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393388 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393400 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393414 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393645 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393660 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.393969 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.394259 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.396395 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.396952 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.398549 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.410159 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.410549 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.410776 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.410822 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.411307 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.411364 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.411516 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.412186 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-56vtp"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.413258 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.415889 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.418024 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.422058 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z274j"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.423035 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.423318 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.423786 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.427272 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.427940 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lbskg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.428298 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.428383 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.428539 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.429179 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.430059 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zlzkl"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.430604 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.431131 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-44d68"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.431711 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.432360 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.433659 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.433821 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.434478 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.434589 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-44xrt"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.435139 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436345 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-config\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436393 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzjrd\" (UniqueName: \"kubernetes.io/projected/27f3af31-4516-4987-9dbd-deb33f197008-kube-api-access-jzjrd\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436420 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46977117-4d75-4653-bb9b-f9f00922f6cb-auth-proxy-config\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436633 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmwrn\" (UniqueName: \"kubernetes.io/projected/5496b467-3482-4ec0-85a0-d01c76d4b522-kube-api-access-vmwrn\") pod \"multus-admission-controller-857f4d67dd-hs4lf\" (UID: \"5496b467-3482-4ec0-85a0-d01c76d4b522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436778 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/60c0644e-874b-42da-9071-09c599cbdf06-tmpfs\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436818 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436834 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqpgf\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-kube-api-access-bqpgf\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436940 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.436983 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-config\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.437017 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-etcd-serving-ca\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.437044 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sf2g\" (UniqueName: \"kubernetes.io/projected/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-kube-api-access-4sf2g\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.437071 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.437095 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e5fb53-aec3-445a-827f-3673c9a18f2b-serving-cert\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.437555 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x4t9b"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.437637 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-etcd-serving-ca\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438242 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438343 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/46977117-4d75-4653-bb9b-f9f00922f6cb-machine-approver-tls\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438391 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq5b8\" (UniqueName: \"kubernetes.io/projected/46977117-4d75-4653-bb9b-f9f00922f6cb-kube-api-access-lq5b8\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438422 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn6dn\" (UniqueName: \"kubernetes.io/projected/7ba0ac21-bc54-430d-a2a7-9703613f38c8-kube-api-access-nn6dn\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438449 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07bf6c8-f1de-4d88-b692-31dec4fe9777-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438476 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-config\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438498 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-encryption-config\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438525 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-serving-cert\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438565 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438592 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-encryption-config\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438621 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07bf6c8-f1de-4d88-b692-31dec4fe9777-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438650 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ea71070-7150-4f2d-b531-08f22697c479-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438675 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-client-ca\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438700 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-audit-policies\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438730 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438748 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438758 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5w55\" (UniqueName: \"kubernetes.io/projected/60c0644e-874b-42da-9071-09c599cbdf06-kube-api-access-p5w55\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438785 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5496b467-3482-4ec0-85a0-d01c76d4b522-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hs4lf\" (UID: \"5496b467-3482-4ec0-85a0-d01c76d4b522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438816 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvpm2\" (UniqueName: \"kubernetes.io/projected/eafbaad4-68ff-47fe-bd96-284913e2e763-kube-api-access-bvpm2\") pod \"migrator-59844c95c7-b2jjx\" (UID: \"eafbaad4-68ff-47fe-bd96-284913e2e763\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438845 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-config\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438871 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438901 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skdzd\" (UniqueName: \"kubernetes.io/projected/b0241b9d-883d-4e59-8f4c-38bf4f8efc91-kube-api-access-skdzd\") pod \"downloads-7954f5f757-25wjt\" (UID: \"b0241b9d-883d-4e59-8f4c-38bf4f8efc91\") " pod="openshift-console/downloads-7954f5f757-25wjt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438924 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-image-import-ca\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438953 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.438980 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-registry-certificates\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439006 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-bound-sa-token\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439033 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ba0ac21-bc54-430d-a2a7-9703613f38c8-profile-collector-cert\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439058 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-client-ca\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439093 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439120 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-etcd-client\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439144 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc55m\" (UniqueName: \"kubernetes.io/projected/b6e5fb53-aec3-445a-827f-3673c9a18f2b-kube-api-access-bc55m\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439168 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439191 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439197 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439215 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f72wv\" (UniqueName: \"kubernetes.io/projected/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-kube-api-access-f72wv\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439243 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ba0ac21-bc54-430d-a2a7-9703613f38c8-srv-cert\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439279 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-node-pullsecrets\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439301 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-audit-dir\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439325 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439349 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htfs2\" (UniqueName: \"kubernetes.io/projected/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-kube-api-access-htfs2\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439387 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-audit\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439412 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-etcd-client\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439437 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-serving-cert\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439469 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-registry-tls\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439494 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60c0644e-874b-42da-9071-09c599cbdf06-webhook-cert\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439524 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439569 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-images\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439604 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ea71070-7150-4f2d-b531-08f22697c479-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439626 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkpgv\" (UniqueName: \"kubernetes.io/projected/37a88100-9cdf-482d-8963-fd6215a3ca67-kube-api-access-rkpgv\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439663 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/27f3af31-4516-4987-9dbd-deb33f197008-audit-dir\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439686 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qxqx\" (UniqueName: \"kubernetes.io/projected/c07bf6c8-f1de-4d88-b692-31dec4fe9777-kube-api-access-8qxqx\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439711 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60c0644e-874b-42da-9071-09c599cbdf06-apiservice-cert\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439737 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46977117-4d75-4653-bb9b-f9f00922f6cb-config\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439759 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37a88100-9cdf-482d-8963-fd6215a3ca67-serving-cert\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439791 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-trusted-ca\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.439809 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-g62qt"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.440018 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.440139 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-config\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.440483 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-audit-policies\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.440969 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xrqgd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.441082 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07bf6c8-f1de-4d88-b692-31dec4fe9777-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.441417 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-node-pullsecrets\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.441439 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-audit-dir\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.441470 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/27f3af31-4516-4987-9dbd-deb33f197008-audit-dir\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.441774 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:25.941758117 +0000 UTC m=+135.435621386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.441976 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-audit\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.443020 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-image-import-ca\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.443438 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.443558 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/27f3af31-4516-4987-9dbd-deb33f197008-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.445883 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07bf6c8-f1de-4d88-b692-31dec4fe9777-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.446407 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-encryption-config\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.446568 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-serving-cert\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.446994 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-serving-cert\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.447165 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5496b467-3482-4ec0-85a0-d01c76d4b522-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hs4lf\" (UID: \"5496b467-3482-4ec0-85a0-d01c76d4b522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.447391 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-encryption-config\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.447615 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/27f3af31-4516-4987-9dbd-deb33f197008-etcd-client\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.448044 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hzqvk"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.448921 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.449680 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hs4lf"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.451377 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2qd7b"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.451815 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.453226 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.453656 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.454752 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.456194 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bchkg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.457500 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-25wjt"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.458919 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kf4mv"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.460230 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.461579 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.463068 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-njg49"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.464473 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-etcd-client\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.465604 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.467864 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.469222 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-56vtp"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.470510 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p67q6"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.475944 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.481500 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.481714 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.481789 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-44d68"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.484612 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.486270 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dnrjc"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.488791 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.490297 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.491144 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.493312 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hzqvk"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.494406 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.494351 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.495321 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.496504 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z274j"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.497586 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zlzkl"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.498773 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x4t9b"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.499815 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lbskg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.501156 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.502485 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.503643 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.504712 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2cc88"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.506152 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-xlwdb"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.506906 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.507055 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.507253 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-44xrt"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.509863 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xlwdb"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.511231 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2cc88"] Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.512821 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.532688 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.540687 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.540843 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.04082091 +0000 UTC m=+135.534684199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.541101 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-config\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.541220 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.541375 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.541497 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-registry-certificates\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.541639 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-bound-sa-token\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.541847 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ba0ac21-bc54-430d-a2a7-9703613f38c8-profile-collector-cert\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542055 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-client-ca\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542182 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542291 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc55m\" (UniqueName: \"kubernetes.io/projected/b6e5fb53-aec3-445a-827f-3673c9a18f2b-kube-api-access-bc55m\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542400 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542519 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f72wv\" (UniqueName: \"kubernetes.io/projected/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-kube-api-access-f72wv\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.542679 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.042655003 +0000 UTC m=+135.536518272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542194 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-config\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542741 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ba0ac21-bc54-430d-a2a7-9703613f38c8-srv-cert\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542816 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542848 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htfs2\" (UniqueName: \"kubernetes.io/projected/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-kube-api-access-htfs2\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542880 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-trusted-ca-bundle\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542905 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vplzj\" (UniqueName: \"kubernetes.io/projected/7bbff2d6-47e4-489b-9c99-b3ef91531447-kube-api-access-vplzj\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542910 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-registry-certificates\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542930 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-client-ca\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.542934 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-registry-tls\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543082 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60c0644e-874b-42da-9071-09c599cbdf06-webhook-cert\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543084 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543188 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543230 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-images\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543260 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ea71070-7150-4f2d-b531-08f22697c479-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543285 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkpgv\" (UniqueName: \"kubernetes.io/projected/37a88100-9cdf-482d-8963-fd6215a3ca67-kube-api-access-rkpgv\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543317 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-service-ca\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543374 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60c0644e-874b-42da-9071-09c599cbdf06-apiservice-cert\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543405 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46977117-4d75-4653-bb9b-f9f00922f6cb-config\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543433 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37a88100-9cdf-482d-8963-fd6215a3ca67-serving-cert\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543471 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-trusted-ca\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543502 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-config\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543569 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-config\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543608 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46977117-4d75-4653-bb9b-f9f00922f6cb-auth-proxy-config\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543680 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/60c0644e-874b-42da-9071-09c599cbdf06-tmpfs\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543710 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqpgf\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-kube-api-access-bqpgf\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543736 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ea71070-7150-4f2d-b531-08f22697c479-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543738 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543796 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-config\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543818 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-serving-cert\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543850 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e5fb53-aec3-445a-827f-3673c9a18f2b-serving-cert\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543868 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/46977117-4d75-4653-bb9b-f9f00922f6cb-machine-approver-tls\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543894 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-oauth-serving-cert\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543928 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq5b8\" (UniqueName: \"kubernetes.io/projected/46977117-4d75-4653-bb9b-f9f00922f6cb-kube-api-access-lq5b8\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543949 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn6dn\" (UniqueName: \"kubernetes.io/projected/7ba0ac21-bc54-430d-a2a7-9703613f38c8-kube-api-access-nn6dn\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543970 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-oauth-config\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.543992 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-client-ca\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.544009 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ea71070-7150-4f2d-b531-08f22697c479-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.544029 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.544046 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5w55\" (UniqueName: \"kubernetes.io/projected/60c0644e-874b-42da-9071-09c599cbdf06-kube-api-access-p5w55\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.544064 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvpm2\" (UniqueName: \"kubernetes.io/projected/eafbaad4-68ff-47fe-bd96-284913e2e763-kube-api-access-bvpm2\") pod \"migrator-59844c95c7-b2jjx\" (UID: \"eafbaad4-68ff-47fe-bd96-284913e2e763\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.544145 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-images\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.544227 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.544808 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.545000 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-config\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.545162 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-trusted-ca\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.545201 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.545236 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/60c0644e-874b-42da-9071-09c599cbdf06-tmpfs\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.545697 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-client-ca\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.546066 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46977117-4d75-4653-bb9b-f9f00922f6cb-config\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.546285 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46977117-4d75-4653-bb9b-f9f00922f6cb-auth-proxy-config\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.546696 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-config\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.547227 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ba0ac21-bc54-430d-a2a7-9703613f38c8-srv-cert\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.548674 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e5fb53-aec3-445a-827f-3673c9a18f2b-serving-cert\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.548940 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-registry-tls\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.549256 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ea71070-7150-4f2d-b531-08f22697c479-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.549287 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60c0644e-874b-42da-9071-09c599cbdf06-apiservice-cert\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.549263 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.549726 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/46977117-4d75-4653-bb9b-f9f00922f6cb-machine-approver-tls\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.549810 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ba0ac21-bc54-430d-a2a7-9703613f38c8-profile-collector-cert\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.550001 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.550135 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37a88100-9cdf-482d-8963-fd6215a3ca67-serving-cert\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.551143 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60c0644e-874b-42da-9071-09c599cbdf06-webhook-cert\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.552371 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.572909 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.613112 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.632755 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645316 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.645471 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.145452003 +0000 UTC m=+135.639315272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645551 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vplzj\" (UniqueName: \"kubernetes.io/projected/7bbff2d6-47e4-489b-9c99-b3ef91531447-kube-api-access-vplzj\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645586 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-trusted-ca-bundle\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645621 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-service-ca\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645687 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-config\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645733 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-serving-cert\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645775 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-oauth-serving-cert\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645813 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-oauth-config\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.645885 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.646147 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.146137444 +0000 UTC m=+135.640000713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.646444 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-config\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.646510 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-service-ca\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.646621 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-trusted-ca-bundle\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.647056 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-oauth-serving-cert\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.650884 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-oauth-config\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.651147 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-serving-cert\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.652505 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.673686 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.699843 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.713504 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.733192 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.746620 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.746847 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.246816333 +0000 UTC m=+135.740679602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.747089 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.747883 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.247874203 +0000 UTC m=+135.741737462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.753431 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.779781 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.792653 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.812948 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.833603 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.848621 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.848968 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.348917884 +0000 UTC m=+135.842781153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.849649 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.850149 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.350120518 +0000 UTC m=+135.843983827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.853032 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.874061 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.893037 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.913114 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.933411 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.950976 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.951089 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.451068396 +0000 UTC m=+135.944931665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.951155 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:25 crc kubenswrapper[4807]: E1205 12:08:25.951819 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.451791627 +0000 UTC m=+135.945654897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.952771 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.973730 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 12:08:25 crc kubenswrapper[4807]: I1205 12:08:25.992370 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.013146 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.032914 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.051697 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.051820 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.551780638 +0000 UTC m=+136.045643907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.051915 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.052325 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.552314674 +0000 UTC m=+136.046177943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.053049 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.072628 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.093429 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.114138 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.133515 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.153274 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.153454 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.653422886 +0000 UTC m=+136.147286175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.153710 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.153809 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.154233 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.654216958 +0000 UTC m=+136.148080227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.173410 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.193300 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.214160 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.233992 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.253819 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.254234 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.254432 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.754403905 +0000 UTC m=+136.248267224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.254758 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.255045 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.755037863 +0000 UTC m=+136.248901132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.274020 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.292986 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.314838 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.333452 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.353591 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.356091 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.356246 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.856229878 +0000 UTC m=+136.350093147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.356416 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.356719 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.856711582 +0000 UTC m=+136.350574851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.372773 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.393367 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.419233 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.432097 4807 request.go:700] Waited for 1.001168163s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.433872 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.453713 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.457023 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.457183 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.957156626 +0000 UTC m=+136.451019885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.457388 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.457715 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:26.957702721 +0000 UTC m=+136.451565990 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.473400 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.493511 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.513891 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.533283 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.553965 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.558232 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.558340 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.05831961 +0000 UTC m=+136.552182889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.558453 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.558776 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.058767902 +0000 UTC m=+136.552631171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.573986 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.593263 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.614163 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.633660 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.653355 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.659546 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.659694 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.159678029 +0000 UTC m=+136.653541298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.659738 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.660071 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.160061331 +0000 UTC m=+136.653924600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.673319 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.700173 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.713627 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.732748 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.752131 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.760672 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.760853 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.260832514 +0000 UTC m=+136.754695793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.761078 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.761388 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.261379719 +0000 UTC m=+136.755242988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.772320 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.792712 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.827741 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzjrd\" (UniqueName: \"kubernetes.io/projected/27f3af31-4516-4987-9dbd-deb33f197008-kube-api-access-jzjrd\") pod \"apiserver-7bbb656c7d-5j622\" (UID: \"27f3af31-4516-4987-9dbd-deb33f197008\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.846744 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmwrn\" (UniqueName: \"kubernetes.io/projected/5496b467-3482-4ec0-85a0-d01c76d4b522-kube-api-access-vmwrn\") pod \"multus-admission-controller-857f4d67dd-hs4lf\" (UID: \"5496b467-3482-4ec0-85a0-d01c76d4b522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.861756 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.861913 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.361878994 +0000 UTC m=+136.855742273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.862147 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.862502 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.362490851 +0000 UTC m=+136.856354120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.867721 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sf2g\" (UniqueName: \"kubernetes.io/projected/156bdc3e-1f65-44cd-8ccd-26bc625a01c5-kube-api-access-4sf2g\") pod \"apiserver-76f77b778f-g62qt\" (UID: \"156bdc3e-1f65-44cd-8ccd-26bc625a01c5\") " pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.873082 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.892883 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.895946 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.913722 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.945201 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.957353 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.963134 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.963505 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.463483621 +0000 UTC m=+136.957346900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.963688 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:26 crc kubenswrapper[4807]: E1205 12:08:26.964004 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.463992136 +0000 UTC m=+136.957855425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.973411 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.979854 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" Dec 05 12:08:26 crc kubenswrapper[4807]: I1205 12:08:26.993652 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.013922 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.034147 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.054383 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.064539 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.064756 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.564731998 +0000 UTC m=+137.058595267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.064858 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.065385 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.565340435 +0000 UTC m=+137.059203704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.072650 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622"] Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.085963 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.092946 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.113646 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.133401 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.134345 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.152511 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.156897 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hs4lf"] Dec 05 12:08:27 crc kubenswrapper[4807]: W1205 12:08:27.163964 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5496b467_3482_4ec0_85a0_d01c76d4b522.slice/crio-2079c043153b3c692010de34a4c7709ebbf65bebefc1edc1364feac6fc573f4b WatchSource:0}: Error finding container 2079c043153b3c692010de34a4c7709ebbf65bebefc1edc1364feac6fc573f4b: Status 404 returned error can't find the container with id 2079c043153b3c692010de34a4c7709ebbf65bebefc1edc1364feac6fc573f4b Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.165410 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.165545 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.665497311 +0000 UTC m=+137.159360580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.165658 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.166000 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.665989575 +0000 UTC m=+137.159852844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.187778 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skdzd\" (UniqueName: \"kubernetes.io/projected/b0241b9d-883d-4e59-8f4c-38bf4f8efc91-kube-api-access-skdzd\") pod \"downloads-7954f5f757-25wjt\" (UID: \"b0241b9d-883d-4e59-8f4c-38bf4f8efc91\") " pod="openshift-console/downloads-7954f5f757-25wjt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.207149 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qxqx\" (UniqueName: \"kubernetes.io/projected/c07bf6c8-f1de-4d88-b692-31dec4fe9777-kube-api-access-8qxqx\") pod \"openshift-apiserver-operator-796bbdcf4f-pw96n\" (UID: \"c07bf6c8-f1de-4d88-b692-31dec4fe9777\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.216817 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.233430 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.253277 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.267009 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.267431 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.767411777 +0000 UTC m=+137.261275046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.273560 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.293599 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-g62qt"] Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.293936 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 12:08:27 crc kubenswrapper[4807]: W1205 12:08:27.300365 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod156bdc3e_1f65_44cd_8ccd_26bc625a01c5.slice/crio-78b0c05101c4cd1104fa9bc8cbc4782ec8932d460735cb26e8c35b99fb809a39 WatchSource:0}: Error finding container 78b0c05101c4cd1104fa9bc8cbc4782ec8932d460735cb26e8c35b99fb809a39: Status 404 returned error can't find the container with id 78b0c05101c4cd1104fa9bc8cbc4782ec8932d460735cb26e8c35b99fb809a39 Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.301911 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-25wjt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.313584 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.325764 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.333186 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.368622 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.368990 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.868975212 +0000 UTC m=+137.362838481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.372614 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.393661 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.413210 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.432238 4807 request.go:700] Waited for 1.924961471s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.434887 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.454766 4807 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.469916 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.470180 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.970143226 +0000 UTC m=+137.464006495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.470267 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.470698 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:27.970681272 +0000 UTC m=+137.464544541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.473434 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.510774 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f8df21c-d9b3-4d94-9e27-dcd36f84acc9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m7fq9\" (UID: \"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.511445 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.527566 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-bound-sa-token\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.551325 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc55m\" (UniqueName: \"kubernetes.io/projected/b6e5fb53-aec3-445a-827f-3673c9a18f2b-kube-api-access-bc55m\") pod \"controller-manager-879f6c89f-h2v8s\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.568212 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f72wv\" (UniqueName: \"kubernetes.io/projected/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-kube-api-access-f72wv\") pod \"marketplace-operator-79b997595-bchkg\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.571944 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.572310 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.072292648 +0000 UTC m=+137.566155917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.589158 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htfs2\" (UniqueName: \"kubernetes.io/projected/b03b24c9-59d5-45cd-8416-7ee0d0b3f08e-kube-api-access-htfs2\") pod \"machine-api-operator-5694c8668f-xrqgd\" (UID: \"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.607705 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvpm2\" (UniqueName: \"kubernetes.io/projected/eafbaad4-68ff-47fe-bd96-284913e2e763-kube-api-access-bvpm2\") pod \"migrator-59844c95c7-b2jjx\" (UID: \"eafbaad4-68ff-47fe-bd96-284913e2e763\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.610751 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.629448 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn6dn\" (UniqueName: \"kubernetes.io/projected/7ba0ac21-bc54-430d-a2a7-9703613f38c8-kube-api-access-nn6dn\") pod \"catalog-operator-68c6474976-jvqdd\" (UID: \"7ba0ac21-bc54-430d-a2a7-9703613f38c8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.651730 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqpgf\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-kube-api-access-bqpgf\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.669591 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq5b8\" (UniqueName: \"kubernetes.io/projected/46977117-4d75-4653-bb9b-f9f00922f6cb-kube-api-access-lq5b8\") pod \"machine-approver-56656f9798-7jd7c\" (UID: \"46977117-4d75-4653-bb9b-f9f00922f6cb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.671466 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.673632 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.673925 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.173913856 +0000 UTC m=+137.667777125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.686991 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5w55\" (UniqueName: \"kubernetes.io/projected/60c0644e-874b-42da-9071-09c599cbdf06-kube-api-access-p5w55\") pod \"packageserver-d55dfcdfc-th6hd\" (UID: \"60c0644e-874b-42da-9071-09c599cbdf06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.708209 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkpgv\" (UniqueName: \"kubernetes.io/projected/37a88100-9cdf-482d-8963-fd6215a3ca67-kube-api-access-rkpgv\") pod \"route-controller-manager-6576b87f9c-gk5fr\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.708327 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.727752 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.747288 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vplzj\" (UniqueName: \"kubernetes.io/projected/7bbff2d6-47e4-489b-9c99-b3ef91531447-kube-api-access-vplzj\") pod \"console-f9d7485db-kf4mv\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.749756 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774086 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.774243 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.274218895 +0000 UTC m=+137.768082194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774420 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774521 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/26646615-02ee-4175-9af9-01403073049b-srv-cert\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774579 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjpvr\" (UniqueName: \"kubernetes.io/projected/a35ba616-01d1-4373-bbe9-84785afa9323-kube-api-access-kjpvr\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774614 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-service-ca-bundle\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774655 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c64561f9-bf38-4a52-be8f-869658009f28-metrics-tls\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774706 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wg5c\" (UniqueName: \"kubernetes.io/projected/26646615-02ee-4175-9af9-01403073049b-kube-api-access-9wg5c\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774746 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-serving-cert\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774773 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-config\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774799 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w287t\" (UniqueName: \"kubernetes.io/projected/753e11b0-5a03-432d-932a-87031ae782ea-kube-api-access-w287t\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774826 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774866 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtgf2\" (UniqueName: \"kubernetes.io/projected/bdc8f24b-e6ac-463c-a917-439ac37b58db-kube-api-access-rtgf2\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774893 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1712044-ceea-44b1-9cbc-c703ae5c656f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774920 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774944 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18c7785f-63c7-4afb-b93c-2fb6639b8d51-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.774974 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpdqs\" (UniqueName: \"kubernetes.io/projected/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-kube-api-access-bpdqs\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775012 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74fc686c-36f9-4cd7-bb33-1845d317a1eb-serving-cert\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775070 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-policies\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775097 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-stats-auth\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775154 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/18c7785f-63c7-4afb-b93c-2fb6639b8d51-images\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775191 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64561f9-bf38-4a52-be8f-869658009f28-trusted-ca\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775215 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-config\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775255 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6jq9\" (UniqueName: \"kubernetes.io/projected/a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6-kube-api-access-t6jq9\") pod \"control-plane-machine-set-operator-78cbb6b69f-r9cjd\" (UID: \"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775288 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftkcs\" (UniqueName: \"kubernetes.io/projected/74fc686c-36f9-4cd7-bb33-1845d317a1eb-kube-api-access-ftkcs\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775317 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-client\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775343 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bfbb6060-35b4-481e-b5c8-81f6a2280af0-metrics-tls\") pod \"dns-operator-744455d44c-44xrt\" (UID: \"bfbb6060-35b4-481e-b5c8-81f6a2280af0\") " pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775374 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64561f9-bf38-4a52-be8f-869658009f28-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775404 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26frt\" (UniqueName: \"kubernetes.io/projected/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-kube-api-access-26frt\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775433 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-service-ca\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775464 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775492 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775548 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5697m\" (UniqueName: \"kubernetes.io/projected/18c7785f-63c7-4afb-b93c-2fb6639b8d51-kube-api-access-5697m\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775602 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj9dz\" (UniqueName: \"kubernetes.io/projected/3e69d864-82d7-4f37-8f11-a7537640e94a-kube-api-access-sj9dz\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775634 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5522a41b-50f2-4ab2-885c-0161017a9ea9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775683 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/38b74738-cfe9-42cc-8c1d-b5c700b18a10-certs\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775713 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775743 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e849f7a7-467d-4124-a342-c8b842e28790-config-volume\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775772 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc3e80de-9521-4aab-9210-0e4e2db9b002-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ckgqz\" (UID: \"dc3e80de-9521-4aab-9210-0e4e2db9b002\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775844 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e69d864-82d7-4f37-8f11-a7537640e94a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775876 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4c2f\" (UniqueName: \"kubernetes.io/projected/8a8686ac-d7c0-440f-a743-db87d2a73b73-kube-api-access-s4c2f\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775910 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35ba616-01d1-4373-bbe9-84785afa9323-config\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775960 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl6m6\" (UniqueName: \"kubernetes.io/projected/c64561f9-bf38-4a52-be8f-869658009f28-kube-api-access-hl6m6\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.775990 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-dir\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776017 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1712044-ceea-44b1-9cbc-c703ae5c656f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776145 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/38b74738-cfe9-42cc-8c1d-b5c700b18a10-node-bootstrap-token\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776186 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18c7785f-63c7-4afb-b93c-2fb6639b8d51-proxy-tls\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776217 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/753e11b0-5a03-432d-932a-87031ae782ea-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776308 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-metrics-certs\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776349 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753e11b0-5a03-432d-932a-87031ae782ea-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776382 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9h6l\" (UniqueName: \"kubernetes.io/projected/dc3e80de-9521-4aab-9210-0e4e2db9b002-kube-api-access-k9h6l\") pod \"package-server-manager-789f6589d5-ckgqz\" (UID: \"dc3e80de-9521-4aab-9210-0e4e2db9b002\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776423 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776474 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776511 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v229\" (UniqueName: \"kubernetes.io/projected/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-kube-api-access-5v229\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776609 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89mkb\" (UniqueName: \"kubernetes.io/projected/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-kube-api-access-89mkb\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776674 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-proxy-tls\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776704 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a35ba616-01d1-4373-bbe9-84785afa9323-trusted-ca\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776770 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-signing-key\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776802 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-service-ca-bundle\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776833 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsbbd\" (UniqueName: \"kubernetes.io/projected/c63ee192-3315-4176-9654-8497d5ba9fd7-kube-api-access-nsbbd\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776868 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/74fc686c-36f9-4cd7-bb33-1845d317a1eb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776898 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-ca\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776926 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v25pg\" (UniqueName: \"kubernetes.io/projected/bfbb6060-35b4-481e-b5c8-81f6a2280af0-kube-api-access-v25pg\") pod \"dns-operator-744455d44c-44xrt\" (UID: \"bfbb6060-35b4-481e-b5c8-81f6a2280af0\") " pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.776971 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-signing-cabundle\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777000 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq5x7\" (UniqueName: \"kubernetes.io/projected/afbcd69b-9f75-4999-af80-31052865b708-kube-api-access-mq5x7\") pod \"ingress-canary-hzqvk\" (UID: \"afbcd69b-9f75-4999-af80-31052865b708\") " pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777031 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777061 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec820b06-dcc4-4180-b22d-019b6696a3b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8269c\" (UID: \"ec820b06-dcc4-4180-b22d-019b6696a3b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777089 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35ba616-01d1-4373-bbe9-84785afa9323-serving-cert\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777122 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777155 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777188 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r9cjd\" (UID: \"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777219 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5522a41b-50f2-4ab2-885c-0161017a9ea9-config\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777250 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777281 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777312 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e849f7a7-467d-4124-a342-c8b842e28790-secret-volume\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777361 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk5x6\" (UniqueName: \"kubernetes.io/projected/ec820b06-dcc4-4180-b22d-019b6696a3b4-kube-api-access-kk5x6\") pod \"cluster-samples-operator-665b6dd947-8269c\" (UID: \"ec820b06-dcc4-4180-b22d-019b6696a3b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777391 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5522a41b-50f2-4ab2-885c-0161017a9ea9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777419 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc8f24b-e6ac-463c-a917-439ac37b58db-serving-cert\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777455 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777512 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-default-certificate\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777575 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e69d864-82d7-4f37-8f11-a7537640e94a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777609 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777639 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777692 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e69d864-82d7-4f37-8f11-a7537640e94a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777727 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-config\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.777760 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngvg5\" (UniqueName: \"kubernetes.io/projected/38b74738-cfe9-42cc-8c1d-b5c700b18a10-kube-api-access-ngvg5\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.778077 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a8686ac-d7c0-440f-a743-db87d2a73b73-config\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.779035 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.279002252 +0000 UTC m=+137.772865521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.779093 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/afbcd69b-9f75-4999-af80-31052865b708-cert\") pod \"ingress-canary-hzqvk\" (UID: \"afbcd69b-9f75-4999-af80-31052865b708\") " pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.779128 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/26646615-02ee-4175-9af9-01403073049b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.779183 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c77ck\" (UniqueName: \"kubernetes.io/projected/e849f7a7-467d-4124-a342-c8b842e28790-kube-api-access-c77ck\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.779423 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a8686ac-d7c0-440f-a743-db87d2a73b73-serving-cert\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.779751 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42sjn\" (UniqueName: \"kubernetes.io/projected/d1712044-ceea-44b1-9cbc-c703ae5c656f-kube-api-access-42sjn\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.833179 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.843305 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880469 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880720 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sz7k\" (UniqueName: \"kubernetes.io/projected/8edc933b-ca85-4e1d-be88-5e619ea40286-kube-api-access-4sz7k\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880765 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/18c7785f-63c7-4afb-b93c-2fb6639b8d51-images\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880805 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64561f9-bf38-4a52-be8f-869658009f28-trusted-ca\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880825 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-config\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880847 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6jq9\" (UniqueName: \"kubernetes.io/projected/a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6-kube-api-access-t6jq9\") pod \"control-plane-machine-set-operator-78cbb6b69f-r9cjd\" (UID: \"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880871 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftkcs\" (UniqueName: \"kubernetes.io/projected/74fc686c-36f9-4cd7-bb33-1845d317a1eb-kube-api-access-ftkcs\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880893 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-client\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880914 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bfbb6060-35b4-481e-b5c8-81f6a2280af0-metrics-tls\") pod \"dns-operator-744455d44c-44xrt\" (UID: \"bfbb6060-35b4-481e-b5c8-81f6a2280af0\") " pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880946 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64561f9-bf38-4a52-be8f-869658009f28-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.880969 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26frt\" (UniqueName: \"kubernetes.io/projected/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-kube-api-access-26frt\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.881007 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.380975169 +0000 UTC m=+137.874838498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881056 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-service-ca\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881099 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881126 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881150 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5697m\" (UniqueName: \"kubernetes.io/projected/18c7785f-63c7-4afb-b93c-2fb6639b8d51-kube-api-access-5697m\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881181 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj9dz\" (UniqueName: \"kubernetes.io/projected/3e69d864-82d7-4f37-8f11-a7537640e94a-kube-api-access-sj9dz\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881203 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5522a41b-50f2-4ab2-885c-0161017a9ea9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881249 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-plugins-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881272 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/38b74738-cfe9-42cc-8c1d-b5c700b18a10-certs\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881296 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881318 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e849f7a7-467d-4124-a342-c8b842e28790-config-volume\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881348 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc3e80de-9521-4aab-9210-0e4e2db9b002-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ckgqz\" (UID: \"dc3e80de-9521-4aab-9210-0e4e2db9b002\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881374 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e69d864-82d7-4f37-8f11-a7537640e94a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881396 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4c2f\" (UniqueName: \"kubernetes.io/projected/8a8686ac-d7c0-440f-a743-db87d2a73b73-kube-api-access-s4c2f\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881422 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35ba616-01d1-4373-bbe9-84785afa9323-config\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881470 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl6m6\" (UniqueName: \"kubernetes.io/projected/c64561f9-bf38-4a52-be8f-869658009f28-kube-api-access-hl6m6\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881553 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/18c7785f-63c7-4afb-b93c-2fb6639b8d51-images\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.881491 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-dir\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882122 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1712044-ceea-44b1-9cbc-c703ae5c656f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882175 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/38b74738-cfe9-42cc-8c1d-b5c700b18a10-node-bootstrap-token\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882199 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18c7785f-63c7-4afb-b93c-2fb6639b8d51-proxy-tls\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882202 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-dir\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882223 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/753e11b0-5a03-432d-932a-87031ae782ea-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882255 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-metrics-certs\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882277 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753e11b0-5a03-432d-932a-87031ae782ea-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882304 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9h6l\" (UniqueName: \"kubernetes.io/projected/dc3e80de-9521-4aab-9210-0e4e2db9b002-kube-api-access-k9h6l\") pod \"package-server-manager-789f6589d5-ckgqz\" (UID: \"dc3e80de-9521-4aab-9210-0e4e2db9b002\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882333 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882360 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-mountpoint-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882389 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882414 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v229\" (UniqueName: \"kubernetes.io/projected/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-kube-api-access-5v229\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882442 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89mkb\" (UniqueName: \"kubernetes.io/projected/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-kube-api-access-89mkb\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882619 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64561f9-bf38-4a52-be8f-869658009f28-trusted-ca\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882659 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-proxy-tls\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882719 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a35ba616-01d1-4373-bbe9-84785afa9323-trusted-ca\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882743 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8edc933b-ca85-4e1d-be88-5e619ea40286-metrics-tls\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882767 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-signing-key\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882790 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-service-ca-bundle\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882878 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsbbd\" (UniqueName: \"kubernetes.io/projected/c63ee192-3315-4176-9654-8497d5ba9fd7-kube-api-access-nsbbd\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882906 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/74fc686c-36f9-4cd7-bb33-1845d317a1eb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882928 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-ca\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.882952 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v25pg\" (UniqueName: \"kubernetes.io/projected/bfbb6060-35b4-481e-b5c8-81f6a2280af0-kube-api-access-v25pg\") pod \"dns-operator-744455d44c-44xrt\" (UID: \"bfbb6060-35b4-481e-b5c8-81f6a2280af0\") " pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.882972 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.382960865 +0000 UTC m=+137.876824244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.883031 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.883035 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-signing-cabundle\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.883307 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35ba616-01d1-4373-bbe9-84785afa9323-config\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.883472 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753e11b0-5a03-432d-932a-87031ae782ea-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.883698 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-service-ca-bundle\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.884170 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-signing-cabundle\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.884499 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/74fc686c-36f9-4cd7-bb33-1845d317a1eb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.885598 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e849f7a7-467d-4124-a342-c8b842e28790-config-volume\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.886093 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq5x7\" (UniqueName: \"kubernetes.io/projected/afbcd69b-9f75-4999-af80-31052865b708-kube-api-access-mq5x7\") pod \"ingress-canary-hzqvk\" (UID: \"afbcd69b-9f75-4999-af80-31052865b708\") " pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.886296 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18c7785f-63c7-4afb-b93c-2fb6639b8d51-proxy-tls\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.886775 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc3e80de-9521-4aab-9210-0e4e2db9b002-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ckgqz\" (UID: \"dc3e80de-9521-4aab-9210-0e4e2db9b002\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.886833 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.886878 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.887364 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec820b06-dcc4-4180-b22d-019b6696a3b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8269c\" (UID: \"ec820b06-dcc4-4180-b22d-019b6696a3b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.887559 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35ba616-01d1-4373-bbe9-84785afa9323-serving-cert\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.887744 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888060 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/38b74738-cfe9-42cc-8c1d-b5c700b18a10-certs\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888118 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888159 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888199 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888289 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r9cjd\" (UID: \"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888327 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5522a41b-50f2-4ab2-885c-0161017a9ea9-config\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888353 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edc933b-ca85-4e1d-be88-5e619ea40286-config-volume\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888375 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888399 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888426 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e849f7a7-467d-4124-a342-c8b842e28790-secret-volume\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888451 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk5x6\" (UniqueName: \"kubernetes.io/projected/ec820b06-dcc4-4180-b22d-019b6696a3b4-kube-api-access-kk5x6\") pod \"cluster-samples-operator-665b6dd947-8269c\" (UID: \"ec820b06-dcc4-4180-b22d-019b6696a3b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888458 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/38b74738-cfe9-42cc-8c1d-b5c700b18a10-node-bootstrap-token\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888472 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5522a41b-50f2-4ab2-885c-0161017a9ea9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888495 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc8f24b-e6ac-463c-a917-439ac37b58db-serving-cert\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888533 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888561 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-default-certificate\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888586 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e69d864-82d7-4f37-8f11-a7537640e94a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888611 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-registration-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888637 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888663 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888689 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e69d864-82d7-4f37-8f11-a7537640e94a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888727 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-config\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888752 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngvg5\" (UniqueName: \"kubernetes.io/projected/38b74738-cfe9-42cc-8c1d-b5c700b18a10-kube-api-access-ngvg5\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888768 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a35ba616-01d1-4373-bbe9-84785afa9323-trusted-ca\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888780 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a8686ac-d7c0-440f-a743-db87d2a73b73-config\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888828 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bfbb6060-35b4-481e-b5c8-81f6a2280af0-metrics-tls\") pod \"dns-operator-744455d44c-44xrt\" (UID: \"bfbb6060-35b4-481e-b5c8-81f6a2280af0\") " pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888833 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/afbcd69b-9f75-4999-af80-31052865b708-cert\") pod \"ingress-canary-hzqvk\" (UID: \"afbcd69b-9f75-4999-af80-31052865b708\") " pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888880 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/26646615-02ee-4175-9af9-01403073049b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888907 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c77ck\" (UniqueName: \"kubernetes.io/projected/e849f7a7-467d-4124-a342-c8b842e28790-kube-api-access-c77ck\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888936 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a8686ac-d7c0-440f-a743-db87d2a73b73-serving-cert\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888960 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-csi-data-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.888982 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gtmc\" (UniqueName: \"kubernetes.io/projected/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-kube-api-access-6gtmc\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889013 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42sjn\" (UniqueName: \"kubernetes.io/projected/d1712044-ceea-44b1-9cbc-c703ae5c656f-kube-api-access-42sjn\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889048 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889079 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/26646615-02ee-4175-9af9-01403073049b-srv-cert\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889102 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjpvr\" (UniqueName: \"kubernetes.io/projected/a35ba616-01d1-4373-bbe9-84785afa9323-kube-api-access-kjpvr\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889102 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889124 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-service-ca-bundle\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889182 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c64561f9-bf38-4a52-be8f-869658009f28-metrics-tls\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889224 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wg5c\" (UniqueName: \"kubernetes.io/projected/26646615-02ee-4175-9af9-01403073049b-kube-api-access-9wg5c\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889264 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-socket-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889312 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-serving-cert\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889347 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-config\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889382 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w287t\" (UniqueName: \"kubernetes.io/projected/753e11b0-5a03-432d-932a-87031ae782ea-kube-api-access-w287t\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889421 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889478 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtgf2\" (UniqueName: \"kubernetes.io/projected/bdc8f24b-e6ac-463c-a917-439ac37b58db-kube-api-access-rtgf2\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889514 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1712044-ceea-44b1-9cbc-c703ae5c656f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889577 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889611 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18c7785f-63c7-4afb-b93c-2fb6639b8d51-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889650 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpdqs\" (UniqueName: \"kubernetes.io/projected/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-kube-api-access-bpdqs\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889718 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74fc686c-36f9-4cd7-bb33-1845d317a1eb-serving-cert\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889773 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-policies\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.889808 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-stats-auth\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.890010 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-service-ca-bundle\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.890361 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.890822 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1712044-ceea-44b1-9cbc-c703ae5c656f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.891318 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.891388 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35ba616-01d1-4373-bbe9-84785afa9323-serving-cert\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.891874 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-proxy-tls\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.891923 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-signing-key\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.891985 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e849f7a7-467d-4124-a342-c8b842e28790-secret-volume\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.891998 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r9cjd\" (UID: \"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.892736 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-policies\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.892968 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.893076 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/afbcd69b-9f75-4999-af80-31052865b708-cert\") pod \"ingress-canary-hzqvk\" (UID: \"afbcd69b-9f75-4999-af80-31052865b708\") " pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.893639 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e69d864-82d7-4f37-8f11-a7537640e94a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.893774 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a8686ac-d7c0-440f-a743-db87d2a73b73-config\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.894583 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a8686ac-d7c0-440f-a743-db87d2a73b73-serving-cert\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.893996 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18c7785f-63c7-4afb-b93c-2fb6639b8d51-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.893782 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-config\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.894268 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec820b06-dcc4-4180-b22d-019b6696a3b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8269c\" (UID: \"ec820b06-dcc4-4180-b22d-019b6696a3b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.894695 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e69d864-82d7-4f37-8f11-a7537640e94a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.895040 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.895596 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1712044-ceea-44b1-9cbc-c703ae5c656f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.896444 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.896508 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/26646615-02ee-4175-9af9-01403073049b-srv-cert\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.896622 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74fc686c-36f9-4cd7-bb33-1845d317a1eb-serving-cert\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.896683 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-config\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.896718 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.897128 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.897268 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/26646615-02ee-4175-9af9-01403073049b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.900513 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.901984 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c64561f9-bf38-4a52-be8f-869658009f28-metrics-tls\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.908560 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6jq9\" (UniqueName: \"kubernetes.io/projected/a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6-kube-api-access-t6jq9\") pod \"control-plane-machine-set-operator-78cbb6b69f-r9cjd\" (UID: \"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.918701 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.929025 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" event={"ID":"156bdc3e-1f65-44cd-8ccd-26bc625a01c5","Type":"ContainerStarted","Data":"78b0c05101c4cd1104fa9bc8cbc4782ec8932d460735cb26e8c35b99fb809a39"} Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.929917 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" event={"ID":"27f3af31-4516-4987-9dbd-deb33f197008","Type":"ContainerStarted","Data":"7eac2b707db23781874cbf048134381fae70550f71d68b35c62fcff2ae27ed5c"} Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.930801 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" event={"ID":"5496b467-3482-4ec0-85a0-d01c76d4b522","Type":"ContainerStarted","Data":"2079c043153b3c692010de34a4c7709ebbf65bebefc1edc1364feac6fc573f4b"} Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.945741 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64561f9-bf38-4a52-be8f-869658009f28-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.977690 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.990926 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991104 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-plugins-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991181 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-mountpoint-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991225 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8edc933b-ca85-4e1d-be88-5e619ea40286-metrics-tls\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991285 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edc933b-ca85-4e1d-be88-5e619ea40286-config-volume\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991332 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-registration-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991373 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-csi-data-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991393 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gtmc\" (UniqueName: \"kubernetes.io/projected/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-kube-api-access-6gtmc\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991425 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-socket-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991486 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sz7k\" (UniqueName: \"kubernetes.io/projected/8edc933b-ca85-4e1d-be88-5e619ea40286-kube-api-access-4sz7k\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991830 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-mountpoint-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: E1205 12:08:27.991941 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.491912132 +0000 UTC m=+137.985775401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991942 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-registration-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.991987 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-csi-data-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.992024 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-socket-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.992248 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-plugins-dir\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.993613 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edc933b-ca85-4e1d-be88-5e619ea40286-config-volume\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:27 crc kubenswrapper[4807]: I1205 12:08:27.996062 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8edc933b-ca85-4e1d-be88-5e619ea40286-metrics-tls\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.005022 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.057394 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-ca\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.057546 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-service-ca\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.062131 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdc8f24b-e6ac-463c-a917-439ac37b58db-config\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.062222 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26frt\" (UniqueName: \"kubernetes.io/projected/6339e59d-8f83-4bb6-9c16-8a95256e1bcf-kube-api-access-26frt\") pod \"service-ca-9c57cc56f-dnrjc\" (UID: \"6339e59d-8f83-4bb6-9c16-8a95256e1bcf\") " pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.062644 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5522a41b-50f2-4ab2-885c-0161017a9ea9-config\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.063958 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5522a41b-50f2-4ab2-885c-0161017a9ea9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.064052 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-stats-auth\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.064607 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.064845 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/753e11b0-5a03-432d-932a-87031ae782ea-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.065178 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5697m\" (UniqueName: \"kubernetes.io/projected/18c7785f-63c7-4afb-b93c-2fb6639b8d51-kube-api-access-5697m\") pod \"machine-config-operator-74547568cd-s8r26\" (UID: \"18c7785f-63c7-4afb-b93c-2fb6639b8d51\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.065274 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-serving-cert\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.065821 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bdc8f24b-e6ac-463c-a917-439ac37b58db-etcd-client\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.068795 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftkcs\" (UniqueName: \"kubernetes.io/projected/74fc686c-36f9-4cd7-bb33-1845d317a1eb-kube-api-access-ftkcs\") pod \"openshift-config-operator-7777fb866f-njg49\" (UID: \"74fc686c-36f9-4cd7-bb33-1845d317a1eb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.069935 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-metrics-certs\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.071240 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-default-certificate\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.071292 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl6m6\" (UniqueName: \"kubernetes.io/projected/c64561f9-bf38-4a52-be8f-869658009f28-kube-api-access-hl6m6\") pod \"ingress-operator-5b745b69d9-mz7vc\" (UID: \"c64561f9-bf38-4a52-be8f-869658009f28\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.081835 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4c2f\" (UniqueName: \"kubernetes.io/projected/8a8686ac-d7c0-440f-a743-db87d2a73b73-kube-api-access-s4c2f\") pod \"service-ca-operator-777779d784-44d68\" (UID: \"8a8686ac-d7c0-440f-a743-db87d2a73b73\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.081842 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj9dz\" (UniqueName: \"kubernetes.io/projected/3e69d864-82d7-4f37-8f11-a7537640e94a-kube-api-access-sj9dz\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.082823 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc8f24b-e6ac-463c-a917-439ac37b58db-serving-cert\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.087805 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e69d864-82d7-4f37-8f11-a7537640e94a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zp9xg\" (UID: \"3e69d864-82d7-4f37-8f11-a7537640e94a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.092641 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.093065 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.593052265 +0000 UTC m=+138.086915534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.097611 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9h6l\" (UniqueName: \"kubernetes.io/projected/dc3e80de-9521-4aab-9210-0e4e2db9b002-kube-api-access-k9h6l\") pod \"package-server-manager-789f6589d5-ckgqz\" (UID: \"dc3e80de-9521-4aab-9210-0e4e2db9b002\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.104373 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.110826 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v25pg\" (UniqueName: \"kubernetes.io/projected/bfbb6060-35b4-481e-b5c8-81f6a2280af0-kube-api-access-v25pg\") pod \"dns-operator-744455d44c-44xrt\" (UID: \"bfbb6060-35b4-481e-b5c8-81f6a2280af0\") " pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.123461 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.130778 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89mkb\" (UniqueName: \"kubernetes.io/projected/ed25959e-c671-4690-a9c7-f2b2a8b67d0b-kube-api-access-89mkb\") pod \"machine-config-controller-84d6567774-z274j\" (UID: \"ed25959e-c671-4690-a9c7-f2b2a8b67d0b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.132137 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.150454 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v229\" (UniqueName: \"kubernetes.io/projected/6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a-kube-api-access-5v229\") pod \"router-default-5444994796-b52pb\" (UID: \"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a\") " pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.169282 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq5x7\" (UniqueName: \"kubernetes.io/projected/afbcd69b-9f75-4999-af80-31052865b708-kube-api-access-mq5x7\") pod \"ingress-canary-hzqvk\" (UID: \"afbcd69b-9f75-4999-af80-31052865b708\") " pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.195117 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.195324 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.69528769 +0000 UTC m=+138.189150969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.196137 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.196712 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.69669512 +0000 UTC m=+138.190558379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: W1205 12:08:28.203149 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46977117_4d75_4653_bb9b_f9f00922f6cb.slice/crio-f6dfe9a9a05795dfada6bb2624ab9b9af4d0d2a7dda594c55129fa62fd5255b5 WatchSource:0}: Error finding container f6dfe9a9a05795dfada6bb2624ab9b9af4d0d2a7dda594c55129fa62fd5255b5: Status 404 returned error can't find the container with id f6dfe9a9a05795dfada6bb2624ab9b9af4d0d2a7dda594c55129fa62fd5255b5 Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.213816 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3d6a1f9-5ac9-40c2-862c-0b879eb91bee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6mxzg\" (UID: \"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.233129 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk5x6\" (UniqueName: \"kubernetes.io/projected/ec820b06-dcc4-4180-b22d-019b6696a3b4-kube-api-access-kk5x6\") pod \"cluster-samples-operator-665b6dd947-8269c\" (UID: \"ec820b06-dcc4-4180-b22d-019b6696a3b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.238183 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.243802 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.255969 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5522a41b-50f2-4ab2-885c-0161017a9ea9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gzz4h\" (UID: \"5522a41b-50f2-4ab2-885c-0161017a9ea9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.275855 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-25wjt"] Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.284358 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsbbd\" (UniqueName: \"kubernetes.io/projected/c63ee192-3315-4176-9654-8497d5ba9fd7-kube-api-access-nsbbd\") pod \"oauth-openshift-558db77b4-x4t9b\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.285943 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.289880 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42sjn\" (UniqueName: \"kubernetes.io/projected/d1712044-ceea-44b1-9cbc-c703ae5c656f-kube-api-access-42sjn\") pod \"openshift-controller-manager-operator-756b6f6bc6-ftv59\" (UID: \"d1712044-ceea-44b1-9cbc-c703ae5c656f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.290447 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.297766 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.298148 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.298369 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.798323568 +0000 UTC m=+138.292186837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.298752 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.299423 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.799392568 +0000 UTC m=+138.293255997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.309305 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c77ck\" (UniqueName: \"kubernetes.io/projected/e849f7a7-467d-4124-a342-c8b842e28790-kube-api-access-c77ck\") pod \"collect-profiles-29415600-xwfzx\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.321643 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.322114 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.324632 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.350949 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.354003 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtgf2\" (UniqueName: \"kubernetes.io/projected/bdc8f24b-e6ac-463c-a917-439ac37b58db-kube-api-access-rtgf2\") pod \"etcd-operator-b45778765-lbskg\" (UID: \"bdc8f24b-e6ac-463c-a917-439ac37b58db\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.354065 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.363236 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.366580 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpdqs\" (UniqueName: \"kubernetes.io/projected/cecd7f96-25b6-4f5e-9a93-c00cf902e00a-kube-api-access-bpdqs\") pod \"authentication-operator-69f744f599-56vtp\" (UID: \"cecd7f96-25b6-4f5e-9a93-c00cf902e00a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.370719 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.372320 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wg5c\" (UniqueName: \"kubernetes.io/projected/26646615-02ee-4175-9af9-01403073049b-kube-api-access-9wg5c\") pod \"olm-operator-6b444d44fb-m9vtg\" (UID: \"26646615-02ee-4175-9af9-01403073049b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.395891 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjpvr\" (UniqueName: \"kubernetes.io/projected/a35ba616-01d1-4373-bbe9-84785afa9323-kube-api-access-kjpvr\") pod \"console-operator-58897d9998-zlzkl\" (UID: \"a35ba616-01d1-4373-bbe9-84785afa9323\") " pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.399646 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.399826 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.89978334 +0000 UTC m=+138.393646609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.399940 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.400299 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:28.900291575 +0000 UTC m=+138.394154844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.412467 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.416100 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w287t\" (UniqueName: \"kubernetes.io/projected/753e11b0-5a03-432d-932a-87031ae782ea-kube-api-access-w287t\") pod \"kube-storage-version-migrator-operator-b67b599dd-kfhwr\" (UID: \"753e11b0-5a03-432d-932a-87031ae782ea\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.440854 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.446420 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngvg5\" (UniqueName: \"kubernetes.io/projected/38b74738-cfe9-42cc-8c1d-b5c700b18a10-kube-api-access-ngvg5\") pod \"machine-config-server-2qd7b\" (UID: \"38b74738-cfe9-42cc-8c1d-b5c700b18a10\") " pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.447461 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hzqvk" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.447722 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kf4mv"] Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.449461 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sz7k\" (UniqueName: \"kubernetes.io/projected/8edc933b-ca85-4e1d-be88-5e619ea40286-kube-api-access-4sz7k\") pod \"dns-default-xlwdb\" (UID: \"8edc933b-ca85-4e1d-be88-5e619ea40286\") " pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.457281 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2qd7b" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.466418 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.471018 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gtmc\" (UniqueName: \"kubernetes.io/projected/1c48a5de-7d23-4f1e-bdee-31d122fd26ad-kube-api-access-6gtmc\") pod \"csi-hostpathplugin-2cc88\" (UID: \"1c48a5de-7d23-4f1e-bdee-31d122fd26ad\") " pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.482585 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" Dec 05 12:08:28 crc kubenswrapper[4807]: W1205 12:08:28.497739 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d9d73ba_2eed_4b7d_925e_47bba9c1ae8a.slice/crio-954518cdfc6fc2ef962a82e5b836478f18792343e0445194e81fb1582efd288a WatchSource:0}: Error finding container 954518cdfc6fc2ef962a82e5b836478f18792343e0445194e81fb1582efd288a: Status 404 returned error can't find the container with id 954518cdfc6fc2ef962a82e5b836478f18792343e0445194e81fb1582efd288a Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.501210 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.501353 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.001334166 +0000 UTC m=+138.495197435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.501718 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.502184 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.002169629 +0000 UTC m=+138.496032898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.602597 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.602867 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.102825089 +0000 UTC m=+138.596688348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.603492 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.603983 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.103966271 +0000 UTC m=+138.597829540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.633244 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.639174 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.686315 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.692824 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.707550 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.707848 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.207833962 +0000 UTC m=+138.701697221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.810775 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.811813 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.311793367 +0000 UTC m=+138.805656636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.912591 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:28 crc kubenswrapper[4807]: E1205 12:08:28.913227 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.413198008 +0000 UTC m=+138.907061277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.933462 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n"] Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.941841 4807 generic.go:334] "Generic (PLEG): container finished" podID="156bdc3e-1f65-44cd-8ccd-26bc625a01c5" containerID="44a6ec1230cb89e7668bd086901c5036ecc6fb87141d40b70989e2a2fc2e925c" exitCode=0 Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.941917 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" event={"ID":"156bdc3e-1f65-44cd-8ccd-26bc625a01c5","Type":"ContainerDied","Data":"44a6ec1230cb89e7668bd086901c5036ecc6fb87141d40b70989e2a2fc2e925c"} Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.950909 4807 generic.go:334] "Generic (PLEG): container finished" podID="27f3af31-4516-4987-9dbd-deb33f197008" containerID="a08eadca1475d888cdfdcff64b98468a6db5b35e560b046ca64be6548bfd1fb6" exitCode=0 Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.950982 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" event={"ID":"27f3af31-4516-4987-9dbd-deb33f197008","Type":"ContainerDied","Data":"a08eadca1475d888cdfdcff64b98468a6db5b35e560b046ca64be6548bfd1fb6"} Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.959628 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" event={"ID":"46977117-4d75-4653-bb9b-f9f00922f6cb","Type":"ContainerStarted","Data":"f6dfe9a9a05795dfada6bb2624ab9b9af4d0d2a7dda594c55129fa62fd5255b5"} Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.962155 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-25wjt" event={"ID":"b0241b9d-883d-4e59-8f4c-38bf4f8efc91","Type":"ContainerStarted","Data":"b0a93a9dad2b7e45269ea2ba44fc788bc469c7b9cb9f8f3bea74d8675937800d"} Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.964927 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-b52pb" event={"ID":"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a","Type":"ContainerStarted","Data":"954518cdfc6fc2ef962a82e5b836478f18792343e0445194e81fb1582efd288a"} Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.984928 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kf4mv" event={"ID":"7bbff2d6-47e4-489b-9c99-b3ef91531447","Type":"ContainerStarted","Data":"3f084501e323d5b722edefe5b3cb87c0fa4f2f3b2af53da08994fcdb46868569"} Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.988152 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd"] Dec 05 12:08:28 crc kubenswrapper[4807]: I1205 12:08:28.994390 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" event={"ID":"5496b467-3482-4ec0-85a0-d01c76d4b522","Type":"ContainerStarted","Data":"c471892bb94d12a2d873a4b150bd3fa999186cb42d41f1fc17432a05e06f647c"} Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.015259 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.021957 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.521936648 +0000 UTC m=+139.015799917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.117047 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.117253 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.617219055 +0000 UTC m=+139.111082334 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.117287 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.117667 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.617658607 +0000 UTC m=+139.111521876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.218949 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.219511 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.719493721 +0000 UTC m=+139.213356990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.284586 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.322669 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.323074 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.823062003 +0000 UTC m=+139.316925282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.324974 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-njg49"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.331631 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xrqgd"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.331691 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.348933 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.424482 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.424963 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:29.924946327 +0000 UTC m=+139.418809586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.461709 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.526249 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:29 crc kubenswrapper[4807]: W1205 12:08:29.537463 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74fc686c_36f9_4cd7_bb33_1845d317a1eb.slice/crio-557bd3d2ae5d388ce10d1a772c80621d3089840d3c9df08c8344f0b32b1c0b68 WatchSource:0}: Error finding container 557bd3d2ae5d388ce10d1a772c80621d3089840d3c9df08c8344f0b32b1c0b68: Status 404 returned error can't find the container with id 557bd3d2ae5d388ce10d1a772c80621d3089840d3c9df08c8344f0b32b1c0b68 Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.577893 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.077848877 +0000 UTC m=+139.571712146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: W1205 12:08:29.602394 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18c7785f_63c7_4afb_b93c_2fb6639b8d51.slice/crio-c29bb1fd0114029ac3e5ac90729060f13cf36dbd52113909ce33d5b9acbaf05b WatchSource:0}: Error finding container c29bb1fd0114029ac3e5ac90729060f13cf36dbd52113909ce33d5b9acbaf05b: Status 404 returned error can't find the container with id c29bb1fd0114029ac3e5ac90729060f13cf36dbd52113909ce33d5b9acbaf05b Dec 05 12:08:29 crc kubenswrapper[4807]: W1205 12:08:29.608558 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeafbaad4_68ff_47fe_bd96_284913e2e763.slice/crio-fdb20d66094d98cc673e84f807b140794541084a75a39598668a2232b582ce35 WatchSource:0}: Error finding container fdb20d66094d98cc673e84f807b140794541084a75a39598668a2232b582ce35: Status 404 returned error can't find the container with id fdb20d66094d98cc673e84f807b140794541084a75a39598668a2232b582ce35 Dec 05 12:08:29 crc kubenswrapper[4807]: W1205 12:08:29.610442 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a88100_9cdf_482d_8963_fd6215a3ca67.slice/crio-29af8102beae524743b0c72ddbca9abb229ce0b42c95356957e851f96419bb71 WatchSource:0}: Error finding container 29af8102beae524743b0c72ddbca9abb229ce0b42c95356957e851f96419bb71: Status 404 returned error can't find the container with id 29af8102beae524743b0c72ddbca9abb229ce0b42c95356957e851f96419bb71 Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.627726 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.628170 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.128138931 +0000 UTC m=+139.622002200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.731425 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.731884 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.231869038 +0000 UTC m=+139.725732307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.756018 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.832393 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.832695 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.332679363 +0000 UTC m=+139.826542632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.832734 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.833041 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.333034992 +0000 UTC m=+139.826898261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.914712 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h2v8s"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.931926 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bchkg"] Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.934055 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:29 crc kubenswrapper[4807]: E1205 12:08:29.934753 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.434731522 +0000 UTC m=+139.928594791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:29 crc kubenswrapper[4807]: I1205 12:08:29.949140 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.009688 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" event={"ID":"71ce4423-09bb-43f1-a7a9-86143ea5dd9d","Type":"ContainerStarted","Data":"23c1bf196b7eed62bc7bfe71c73f9ee97306930633c6551bf853a8117aab78e4"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.012753 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" event={"ID":"5496b467-3482-4ec0-85a0-d01c76d4b522","Type":"ContainerStarted","Data":"6469c19631f31b09adc52953155577f2c730effed760bbce46de11b7257074fb"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.015913 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" event={"ID":"18c7785f-63c7-4afb-b93c-2fb6639b8d51","Type":"ContainerStarted","Data":"05de7dcb74989db6ac76f8b0a5ab03a9f344ae12c6a5f16c5ae3cfc65ea59456"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.015957 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" event={"ID":"18c7785f-63c7-4afb-b93c-2fb6639b8d51","Type":"ContainerStarted","Data":"c29bb1fd0114029ac3e5ac90729060f13cf36dbd52113909ce33d5b9acbaf05b"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.020382 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2qd7b" event={"ID":"38b74738-cfe9-42cc-8c1d-b5c700b18a10","Type":"ContainerStarted","Data":"b4fcc578040fe9017efb97a439a3dc627513a2e18ff84840fce9228b6590da3d"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.020429 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2qd7b" event={"ID":"38b74738-cfe9-42cc-8c1d-b5c700b18a10","Type":"ContainerStarted","Data":"ce8cc8bfaa614627dfb0f6142b556853ac4bb9b6d4079b932b91dd5c46531ac6"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.023035 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" event={"ID":"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e","Type":"ContainerStarted","Data":"a8085b219b56cbade10146674299f6484ccb9d740366e7bd7367a1e785fa178d"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.023067 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" event={"ID":"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e","Type":"ContainerStarted","Data":"e84b598cd35a7e7efe3862ed7506fa17dc9ebb7255bdf92b2ad8a4faa70ee3c5"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.024035 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" event={"ID":"b6e5fb53-aec3-445a-827f-3673c9a18f2b","Type":"ContainerStarted","Data":"1f776d94f3da1771d062b22a54d114f0e9c185396a8d135204af5a607cf855eb"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.034056 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-hs4lf" podStartSLOduration=119.034041003 podStartE2EDuration="1m59.034041003s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.03220124 +0000 UTC m=+139.526064509" watchObservedRunningTime="2025-12-05 12:08:30.034041003 +0000 UTC m=+139.527904272" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.036514 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.036823 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.536811512 +0000 UTC m=+140.030674781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.042728 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-25wjt" event={"ID":"b0241b9d-883d-4e59-8f4c-38bf4f8efc91","Type":"ContainerStarted","Data":"ce4132175455620770e4d748b558857e133a97e2fa7d3a349f45a4b986672de3"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.047696 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-25wjt" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.056857 4807 patch_prober.go:28] interesting pod/downloads-7954f5f757-25wjt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.056923 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25wjt" podUID="b0241b9d-883d-4e59-8f4c-38bf4f8efc91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.063309 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kf4mv" event={"ID":"7bbff2d6-47e4-489b-9c99-b3ef91531447","Type":"ContainerStarted","Data":"b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.081828 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2qd7b" podStartSLOduration=5.081809415 podStartE2EDuration="5.081809415s" podCreationTimestamp="2025-12-05 12:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.048507716 +0000 UTC m=+139.542370975" watchObservedRunningTime="2025-12-05 12:08:30.081809415 +0000 UTC m=+139.575672684" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.082201 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-25wjt" podStartSLOduration=119.082197236 podStartE2EDuration="1m59.082197236s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.076912105 +0000 UTC m=+139.570775394" watchObservedRunningTime="2025-12-05 12:08:30.082197236 +0000 UTC m=+139.576060495" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.098904 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-b52pb" event={"ID":"6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a","Type":"ContainerStarted","Data":"9a078ef7fb77145e5030263df8df2eba047ca276336adba21c49131b7ef7a3f6"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.100865 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.105999 4807 generic.go:334] "Generic (PLEG): container finished" podID="74fc686c-36f9-4cd7-bb33-1845d317a1eb" containerID="f0f2fc56ee6dbd7cdf5f8a91fc00311955070610bf56de1cea0b254b285be7d3" exitCode=0 Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.106111 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" event={"ID":"74fc686c-36f9-4cd7-bb33-1845d317a1eb","Type":"ContainerDied","Data":"f0f2fc56ee6dbd7cdf5f8a91fc00311955070610bf56de1cea0b254b285be7d3"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.106110 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kf4mv" podStartSLOduration=119.106090567 podStartE2EDuration="1m59.106090567s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.104664757 +0000 UTC m=+139.598528026" watchObservedRunningTime="2025-12-05 12:08:30.106090567 +0000 UTC m=+139.599953826" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.106140 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" event={"ID":"74fc686c-36f9-4cd7-bb33-1845d317a1eb","Type":"ContainerStarted","Data":"557bd3d2ae5d388ce10d1a772c80621d3089840d3c9df08c8344f0b32b1c0b68"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.110252 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" event={"ID":"60c0644e-874b-42da-9071-09c599cbdf06","Type":"ContainerStarted","Data":"13cba2258d0c1f2178331dbd34998dc3e0473a6eec45acfc48c506fdbc88212d"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.112410 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" event={"ID":"46977117-4d75-4653-bb9b-f9f00922f6cb","Type":"ContainerStarted","Data":"a00afe02857c45f9f14b98041e7c262c3f7e9591365a92644da80bce787c5a02"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.113305 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" event={"ID":"37a88100-9cdf-482d-8963-fd6215a3ca67","Type":"ContainerStarted","Data":"29af8102beae524743b0c72ddbca9abb229ce0b42c95356957e851f96419bb71"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.114322 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.115764 4807 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-gk5fr container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.115804 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" podUID="37a88100-9cdf-482d-8963-fd6215a3ca67" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.117209 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" event={"ID":"7ba0ac21-bc54-430d-a2a7-9703613f38c8","Type":"ContainerStarted","Data":"99adb78bf09b77037e030f09f453fff1487c36d8967fdd6f0241ca0d33c0d5bd"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.117240 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" event={"ID":"7ba0ac21-bc54-430d-a2a7-9703613f38c8","Type":"ContainerStarted","Data":"20dbc9780c60868b3a62f474a2fbf2dfed66e871bca65c2d2ffe7cbcf294b958"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.117887 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.119241 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" event={"ID":"c07bf6c8-f1de-4d88-b692-31dec4fe9777","Type":"ContainerStarted","Data":"99479413f2882172a178f70827b830473c3289095b353ffadfdbd58ae3da2582"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.119268 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" event={"ID":"c07bf6c8-f1de-4d88-b692-31dec4fe9777","Type":"ContainerStarted","Data":"56a10fcf580435c357aefdfea9b5b873a5f03c5b682ad8d7a59363f2a61c7d61"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.120635 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" event={"ID":"eafbaad4-68ff-47fe-bd96-284913e2e763","Type":"ContainerStarted","Data":"3a2dbaef507e815d6f670777577df180de1fe209cdeea647f6ebd2816c8ae65e"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.120668 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" event={"ID":"eafbaad4-68ff-47fe-bd96-284913e2e763","Type":"ContainerStarted","Data":"fdb20d66094d98cc673e84f807b140794541084a75a39598668a2232b582ce35"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.123357 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" event={"ID":"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9","Type":"ContainerStarted","Data":"2c3697aff00d5eeb479a782d9c025f3e761ad1f56cdff6a5b41b61d2b2f861ac"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.128069 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-b52pb" podStartSLOduration=119.128055374 podStartE2EDuration="1m59.128055374s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.126647444 +0000 UTC m=+139.620510713" watchObservedRunningTime="2025-12-05 12:08:30.128055374 +0000 UTC m=+139.621918643" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.129309 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" event={"ID":"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6","Type":"ContainerStarted","Data":"a2834b7c0922180f39b983b30b6f531a301b4407995e11309afac3a26824ad11"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.129341 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" event={"ID":"a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6","Type":"ContainerStarted","Data":"3ebdb4084c0b6ce4900a773e4894382d4dd7b043191a4209786462929824944c"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.132094 4807 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jvqdd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.132129 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" podUID="7ba0ac21-bc54-430d-a2a7-9703613f38c8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.133678 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" event={"ID":"27f3af31-4516-4987-9dbd-deb33f197008","Type":"ContainerStarted","Data":"a63a3f8ed6ac8d648e69a304e24fccbb009840a76206b810cb63b08911ced5c2"} Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.144583 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.145598 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.645582203 +0000 UTC m=+140.139445472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.211158 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" podStartSLOduration=119.211097231 podStartE2EDuration="1m59.211097231s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.208950679 +0000 UTC m=+139.702813958" watchObservedRunningTime="2025-12-05 12:08:30.211097231 +0000 UTC m=+139.704960500" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.257106 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hzqvk"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.262062 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.266898 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.766880892 +0000 UTC m=+140.260744151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.291147 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.305681 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59"] Dec 05 12:08:30 crc kubenswrapper[4807]: W1205 12:08:30.294835 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3d6a1f9_5ac9_40c2_862c_0b879eb91bee.slice/crio-905082ed3559e9afde803df8a9f8d4d3aefdc500f1b6ab819f9825868cb5986f WatchSource:0}: Error finding container 905082ed3559e9afde803df8a9f8d4d3aefdc500f1b6ab819f9825868cb5986f: Status 404 returned error can't find the container with id 905082ed3559e9afde803df8a9f8d4d3aefdc500f1b6ab819f9825868cb5986f Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.299826 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:30 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:30 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:30 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.305857 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.306696 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.315837 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.317861 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dnrjc"] Dec 05 12:08:30 crc kubenswrapper[4807]: W1205 12:08:30.328026 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafbcd69b_9f75_4999_af80_31052865b708.slice/crio-4579b4f04bc78ec41447354a3d191154bb78b37c0a99cb0907f858861fe5f220 WatchSource:0}: Error finding container 4579b4f04bc78ec41447354a3d191154bb78b37c0a99cb0907f858861fe5f220: Status 404 returned error can't find the container with id 4579b4f04bc78ec41447354a3d191154bb78b37c0a99cb0907f858861fe5f220 Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.328200 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.347260 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-44xrt"] Dec 05 12:08:30 crc kubenswrapper[4807]: W1205 12:08:30.355910 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded25959e_c671_4690_a9c7_f2b2a8b67d0b.slice/crio-18ce7c7357fc9bc117dd46f9fcf1ef917a9ab36d225594cdff24b1c4d548d2c1 WatchSource:0}: Error finding container 18ce7c7357fc9bc117dd46f9fcf1ef917a9ab36d225594cdff24b1c4d548d2c1: Status 404 returned error can't find the container with id 18ce7c7357fc9bc117dd46f9fcf1ef917a9ab36d225594cdff24b1c4d548d2c1 Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.358879 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-56vtp"] Dec 05 12:08:30 crc kubenswrapper[4807]: W1205 12:08:30.363958 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecd7f96_25b6_4f5e_9a93_c00cf902e00a.slice/crio-c0a088fa55d0889800806adb68af5c722a3165b193cd87e2a84a62a7634290b7 WatchSource:0}: Error finding container c0a088fa55d0889800806adb68af5c722a3165b193cd87e2a84a62a7634290b7: Status 404 returned error can't find the container with id c0a088fa55d0889800806adb68af5c722a3165b193cd87e2a84a62a7634290b7 Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.364697 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.365031 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.865015419 +0000 UTC m=+140.358878688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.367197 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xlwdb"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.372976 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" podStartSLOduration=119.372954476 podStartE2EDuration="1m59.372954476s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.257941856 +0000 UTC m=+139.751805125" watchObservedRunningTime="2025-12-05 12:08:30.372954476 +0000 UTC m=+139.866817755" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.388284 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.391512 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.393786 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lbskg"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.400950 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.409853 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-44d68"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.418693 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z274j"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.419976 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw96n" podStartSLOduration=120.419919574 podStartE2EDuration="2m0.419919574s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.291791172 +0000 UTC m=+139.785654441" watchObservedRunningTime="2025-12-05 12:08:30.419919574 +0000 UTC m=+139.913782863" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.431779 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" podStartSLOduration=119.431754972 podStartE2EDuration="1m59.431754972s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.369743354 +0000 UTC m=+139.863606643" watchObservedRunningTime="2025-12-05 12:08:30.431754972 +0000 UTC m=+139.925618241" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.436618 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r9cjd" podStartSLOduration=119.43660041 podStartE2EDuration="1m59.43660041s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:30.405447532 +0000 UTC m=+139.899310801" watchObservedRunningTime="2025-12-05 12:08:30.43660041 +0000 UTC m=+139.930463679" Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.451912 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.455874 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x4t9b"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.458475 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zlzkl"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.464560 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2cc88"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.465365 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr"] Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.466603 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.467445 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:30.967420179 +0000 UTC m=+140.461283448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: W1205 12:08:30.486094 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda35ba616_01d1_4373_bbe9_84785afa9323.slice/crio-c0014147debbaa2660a0989c99a672c2bdcf904fd76d724182baf41d256a5cae WatchSource:0}: Error finding container c0014147debbaa2660a0989c99a672c2bdcf904fd76d724182baf41d256a5cae: Status 404 returned error can't find the container with id c0014147debbaa2660a0989c99a672c2bdcf904fd76d724182baf41d256a5cae Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.568298 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.568492 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.068460989 +0000 UTC m=+140.562324258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.568806 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.569366 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.069339975 +0000 UTC m=+140.563203254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.674416 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.675348 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.175326276 +0000 UTC m=+140.669189555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.785639 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.786424 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.286409363 +0000 UTC m=+140.780272632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.890425 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.890671 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.390634954 +0000 UTC m=+140.884498233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.891003 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.891578 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.391559901 +0000 UTC m=+140.885423170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.997608 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.997859 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.4977897 +0000 UTC m=+140.991652979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:30 crc kubenswrapper[4807]: I1205 12:08:30.998385 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:30 crc kubenswrapper[4807]: E1205 12:08:30.999278 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.499256311 +0000 UTC m=+140.993119590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.118716 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.119507 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.61949084 +0000 UTC m=+141.113354109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.188681 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" event={"ID":"46977117-4d75-4653-bb9b-f9f00922f6cb","Type":"ContainerStarted","Data":"2ab3e09516d4e3a95564706b3f51efd241bb691b20775e8e4fc34ef1ccc4fbef"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.194304 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" event={"ID":"a35ba616-01d1-4373-bbe9-84785afa9323","Type":"ContainerStarted","Data":"c0014147debbaa2660a0989c99a672c2bdcf904fd76d724182baf41d256a5cae"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.195494 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" event={"ID":"5522a41b-50f2-4ab2-885c-0161017a9ea9","Type":"ContainerStarted","Data":"ad61746e67a4e3187e4cfcd161622d9022a3151e28120687dbd5a744506383f5"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.199043 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" event={"ID":"3e69d864-82d7-4f37-8f11-a7537640e94a","Type":"ContainerStarted","Data":"98bf40c9967c99d4e57fd1eaf48be5ef3b60c14c9c4ef3a5ed1cdd4cbecf83d0"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.210679 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" event={"ID":"b6e5fb53-aec3-445a-827f-3673c9a18f2b","Type":"ContainerStarted","Data":"0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.211954 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.215480 4807 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-h2v8s container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.215538 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" podUID="b6e5fb53-aec3-445a-827f-3673c9a18f2b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.218869 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7jd7c" podStartSLOduration=122.218848052 podStartE2EDuration="2m2.218848052s" podCreationTimestamp="2025-12-05 12:06:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:31.211926835 +0000 UTC m=+140.705790124" watchObservedRunningTime="2025-12-05 12:08:31.218848052 +0000 UTC m=+140.712711321" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.221708 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.222030 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.722018502 +0000 UTC m=+141.215881771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.228146 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" event={"ID":"bdc8f24b-e6ac-463c-a917-439ac37b58db","Type":"ContainerStarted","Data":"cd353639914e1d838a8569f8d82298e6d590c990a774c4ca9e5cc287541c8008"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.228188 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" event={"ID":"bdc8f24b-e6ac-463c-a917-439ac37b58db","Type":"ContainerStarted","Data":"3883e808c72a3f93674826363724f57e4a6cedcd1779952ed2db0adb8f82f752"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.264743 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" event={"ID":"b03b24c9-59d5-45cd-8416-7ee0d0b3f08e","Type":"ContainerStarted","Data":"c2075d6627769c9154cb73f876145f19ca44f04871bdf79ef05f0d993ba37cf5"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.273253 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" podStartSLOduration=120.273236723 podStartE2EDuration="2m0.273236723s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:31.242009822 +0000 UTC m=+140.735873091" watchObservedRunningTime="2025-12-05 12:08:31.273236723 +0000 UTC m=+140.767099992" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.289403 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:31 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:31 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:31 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.289465 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.290842 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" event={"ID":"d1712044-ceea-44b1-9cbc-c703ae5c656f","Type":"ContainerStarted","Data":"96cd64e9b69f184e7efe8872b99de27c8af80c479dbb272ea9951bf0bfb88ab4"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.290893 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" event={"ID":"d1712044-ceea-44b1-9cbc-c703ae5c656f","Type":"ContainerStarted","Data":"698b0a32fe31be82b546f81eb2b62bc108d03fe4545e08949b4a910eb28566f4"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.304851 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" event={"ID":"6339e59d-8f83-4bb6-9c16-8a95256e1bcf","Type":"ContainerStarted","Data":"62750b95bc6d370139a5d0babf8323e12c5e95299f4fbe6cda9ca4d5177962e0"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.304896 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" event={"ID":"6339e59d-8f83-4bb6-9c16-8a95256e1bcf","Type":"ContainerStarted","Data":"b626ebab9b98a1123f7f16b2d01d4547ff8f70208e87d7e92f3296c47f57a7d8"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.311711 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" event={"ID":"74fc686c-36f9-4cd7-bb33-1845d317a1eb","Type":"ContainerStarted","Data":"35ae3059d58bf7ee0f9a5874a22d08425532d6e76d0262b765f4ae4528b0e6ff"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.322494 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.324224 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.824198946 +0000 UTC m=+141.318062275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.342174 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" event={"ID":"ed25959e-c671-4690-a9c7-f2b2a8b67d0b","Type":"ContainerStarted","Data":"18ce7c7357fc9bc117dd46f9fcf1ef917a9ab36d225594cdff24b1c4d548d2c1"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.354622 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" event={"ID":"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee","Type":"ContainerStarted","Data":"905082ed3559e9afde803df8a9f8d4d3aefdc500f1b6ab819f9825868cb5986f"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.358226 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-lbskg" podStartSLOduration=120.358204785 podStartE2EDuration="2m0.358204785s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:31.274450097 +0000 UTC m=+140.768313366" watchObservedRunningTime="2025-12-05 12:08:31.358204785 +0000 UTC m=+140.852068044" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.386440 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xlwdb" event={"ID":"8edc933b-ca85-4e1d-be88-5e619ea40286","Type":"ContainerStarted","Data":"4b23416cf638046a006a02f8202af43ee4893edb9bc814a407613e8b38cb5f50"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.388269 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-xrqgd" podStartSLOduration=120.388248042 podStartE2EDuration="2m0.388248042s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:31.386881463 +0000 UTC m=+140.880744722" watchObservedRunningTime="2025-12-05 12:08:31.388248042 +0000 UTC m=+140.882111311" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.395046 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" event={"ID":"26646615-02ee-4175-9af9-01403073049b","Type":"ContainerStarted","Data":"c6105f8e1ee0c3c3ea0cde11b27467510f84c52d5d6f0843dfccf2e94889a930"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.426679 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.429130 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:31.929116437 +0000 UTC m=+141.422979706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.436179 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" event={"ID":"8a8686ac-d7c0-440f-a743-db87d2a73b73","Type":"ContainerStarted","Data":"420e7a6a513ab39fec8871b7ca5a0571824b31ee6398a02a4b23b327ae7ec680"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.437792 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" event={"ID":"8a8686ac-d7c0-440f-a743-db87d2a73b73","Type":"ContainerStarted","Data":"08333081a7f5f08ef123b44e8d6ed692deab8b79a285a0f928aa02f3311d2bf4"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.496579 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" event={"ID":"dc3e80de-9521-4aab-9210-0e4e2db9b002","Type":"ContainerStarted","Data":"f1b0d336b7749e9ff3deb0705bfa8c48cc28123dd8d417ab6d4edfa40b36f0a5"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.496633 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" event={"ID":"dc3e80de-9521-4aab-9210-0e4e2db9b002","Type":"ContainerStarted","Data":"bf5768020c9a4e90bd0b370327a3989aa34509c60ca92a709b3746b081a520a6"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.499084 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" event={"ID":"18c7785f-63c7-4afb-b93c-2fb6639b8d51","Type":"ContainerStarted","Data":"5918b508668693e1197a911d8f60e53683fe2bcddb4424a991a1decb03de378d"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.504847 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" event={"ID":"156bdc3e-1f65-44cd-8ccd-26bc625a01c5","Type":"ContainerStarted","Data":"de3697c0dbbc9b9f17521185d411eec9f65f8783208d0898d7f95f3cb8d59a6c"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.505866 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" event={"ID":"c63ee192-3315-4176-9654-8497d5ba9fd7","Type":"ContainerStarted","Data":"ca41fbc46910ffcaee583675c6bb391edae416ecdf5e0c39957194a32b2a477b"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.539271 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.551197 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.051171117 +0000 UTC m=+141.545034386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.592069 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" event={"ID":"3f8df21c-d9b3-4d94-9e27-dcd36f84acc9","Type":"ContainerStarted","Data":"1d2da1354d253d5108fba200ce03248077c9b17c9d1058493b7649fb2e7dae88"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.667880 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.670194 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.170170269 +0000 UTC m=+141.664033538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.754842 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" event={"ID":"e849f7a7-467d-4124-a342-c8b842e28790","Type":"ContainerStarted","Data":"732ee5c2c7abbe8742e8497bd91870d87398bf1b0adc87585b2ae5f7d2f6d0cd"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.758254 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" event={"ID":"753e11b0-5a03-432d-932a-87031ae782ea","Type":"ContainerStarted","Data":"0192b6a138c27768015101c6700ddee5f2ca39b9e41e50dbc1d6d3597684e377"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.771502 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.772546 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.272505457 +0000 UTC m=+141.766368726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.793926 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" event={"ID":"eafbaad4-68ff-47fe-bd96-284913e2e763","Type":"ContainerStarted","Data":"c4cf59479df84ed7480b00226ae790ed8aa07142ee0928f908aa6db2db215afc"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.837703 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hzqvk" event={"ID":"afbcd69b-9f75-4999-af80-31052865b708","Type":"ContainerStarted","Data":"4579b4f04bc78ec41447354a3d191154bb78b37c0a99cb0907f858861fe5f220"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.873241 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.873586 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.373573268 +0000 UTC m=+141.867436537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.898618 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.899877 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.911374 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" event={"ID":"71ce4423-09bb-43f1-a7a9-86143ea5dd9d","Type":"ContainerStarted","Data":"f4ab674188cc0f8873a0882d7a969dc85235e8c5d9409617ec7dfdda9c4a1b16"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.912188 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.913387 4807 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bchkg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/healthz\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.913473 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.15:8080/healthz\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.954731 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" event={"ID":"bfbb6060-35b4-481e-b5c8-81f6a2280af0","Type":"ContainerStarted","Data":"e6efba5b052f8dc0a96855e10dc00fcf8e1756e9c69f8dab358a70a45c0c3b24"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.956678 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" event={"ID":"60c0644e-874b-42da-9071-09c599cbdf06","Type":"ContainerStarted","Data":"0153fca665690ace07823d6a25096ed0b72883a4a69f85ab23a48cbb27e63cc0"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.957545 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.977006 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.977375 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.477332896 +0000 UTC m=+141.971196165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.978119 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:31 crc kubenswrapper[4807]: E1205 12:08:31.980245 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.480221858 +0000 UTC m=+141.974085177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.983091 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" event={"ID":"cecd7f96-25b6-4f5e-9a93-c00cf902e00a","Type":"ContainerStarted","Data":"c0a088fa55d0889800806adb68af5c722a3165b193cd87e2a84a62a7634290b7"} Dec 05 12:08:31 crc kubenswrapper[4807]: I1205 12:08:31.995668 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" event={"ID":"ec820b06-dcc4-4180-b22d-019b6696a3b4","Type":"ContainerStarted","Data":"8a8636e372cad7aa55cc1a52f41b784ff1a9d7533828b75f28e6f677a9d897ef"} Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.034957 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" event={"ID":"c64561f9-bf38-4a52-be8f-869658009f28","Type":"ContainerStarted","Data":"cbeee5e6f5f0324ed95e6b8790a21d939e0eed07db88dc689d61d75b20d684eb"} Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.079626 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.080952 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.5809375 +0000 UTC m=+142.074800769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.088847 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" event={"ID":"37a88100-9cdf-482d-8963-fd6215a3ca67","Type":"ContainerStarted","Data":"f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f"} Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.111964 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.113331 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" event={"ID":"1c48a5de-7d23-4f1e-bdee-31d122fd26ad","Type":"ContainerStarted","Data":"18142a25f51cac35f477aeee785aa6808fba90f889739f4ca1cc54db715f10e4"} Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.114011 4807 patch_prober.go:28] interesting pod/downloads-7954f5f757-25wjt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.114052 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25wjt" podUID="b0241b9d-883d-4e59-8f4c-38bf4f8efc91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.133992 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jvqdd" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.187010 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.188653 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.6886363 +0000 UTC m=+142.182499569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.288227 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.288622 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.78857704 +0000 UTC m=+142.282440299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.288863 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.290059 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.790043742 +0000 UTC m=+142.283907011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.297579 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:32 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:32 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:32 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.297633 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.390100 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.390782 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.890743003 +0000 UTC m=+142.384606272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.404121 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4d4jf"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.405347 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.420247 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.441593 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4d4jf"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.498825 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-utilities\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.498876 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-catalog-content\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.498903 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfj9l\" (UniqueName: \"kubernetes.io/projected/c9af52ad-e3d9-426f-9adf-a9880437d83f-kube-api-access-zfj9l\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.498986 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.499294 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:32.999278057 +0000 UTC m=+142.493141326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.524143 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-927f9"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.526253 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.548277 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.556107 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-927f9"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.600043 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.600692 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-utilities\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.600799 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-catalog-content\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.600903 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfj9l\" (UniqueName: \"kubernetes.io/projected/c9af52ad-e3d9-426f-9adf-a9880437d83f-kube-api-access-zfj9l\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.601384 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.101368418 +0000 UTC m=+142.595231687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.601816 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-utilities\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.602086 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-catalog-content\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.652392 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfj9l\" (UniqueName: \"kubernetes.io/projected/c9af52ad-e3d9-426f-9adf-a9880437d83f-kube-api-access-zfj9l\") pod \"community-operators-4d4jf\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.654853 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" podStartSLOduration=121.654835712 podStartE2EDuration="2m1.654835712s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:32.654110902 +0000 UTC m=+142.147974171" watchObservedRunningTime="2025-12-05 12:08:32.654835712 +0000 UTC m=+142.148698981" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.704988 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.705100 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t4gl\" (UniqueName: \"kubernetes.io/projected/55077615-de5e-48c6-9be0-2f3c8bbabae7-kube-api-access-5t4gl\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.705134 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-utilities\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.705200 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-catalog-content\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.705280 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.2052679 +0000 UTC m=+142.699131169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.722019 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ff68x"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.723158 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.740326 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ff68x"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.740612 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.765384 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s8r26" podStartSLOduration=121.765366543 podStartE2EDuration="2m1.765366543s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:32.763364356 +0000 UTC m=+142.257227625" watchObservedRunningTime="2025-12-05 12:08:32.765366543 +0000 UTC m=+142.259229812" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.767200 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.866819 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.867455 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.367436833 +0000 UTC m=+142.861300102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.867489 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-catalog-content\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.867565 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x7xv\" (UniqueName: \"kubernetes.io/projected/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-kube-api-access-7x7xv\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.867615 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.867678 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-catalog-content\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.867729 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t4gl\" (UniqueName: \"kubernetes.io/projected/55077615-de5e-48c6-9be0-2f3c8bbabae7-kube-api-access-5t4gl\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.867763 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-utilities\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.867788 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-utilities\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.868232 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-utilities\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.868934 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.368917105 +0000 UTC m=+142.862780374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.868474 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-catalog-content\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.878782 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" podStartSLOduration=122.878764646 podStartE2EDuration="2m2.878764646s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:32.808892254 +0000 UTC m=+142.302755523" watchObservedRunningTime="2025-12-05 12:08:32.878764646 +0000 UTC m=+142.372627915" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.878905 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-th6hd" podStartSLOduration=121.8789005 podStartE2EDuration="2m1.8789005s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:32.876715288 +0000 UTC m=+142.370578557" watchObservedRunningTime="2025-12-05 12:08:32.8789005 +0000 UTC m=+142.372763769" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.906299 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sb9pm"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.907682 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.922279 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t4gl\" (UniqueName: \"kubernetes.io/projected/55077615-de5e-48c6-9be0-2f3c8bbabae7-kube-api-access-5t4gl\") pod \"certified-operators-927f9\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.923964 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.947539 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sb9pm"] Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973071 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973216 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljl7l\" (UniqueName: \"kubernetes.io/projected/792cf8c2-f4b4-4956-955d-43541a2adcd3-kube-api-access-ljl7l\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973244 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-utilities\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973287 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-utilities\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.973353 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.473325533 +0000 UTC m=+142.967188802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973486 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x7xv\" (UniqueName: \"kubernetes.io/projected/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-kube-api-access-7x7xv\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973567 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-catalog-content\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973608 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973670 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-catalog-content\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.973692 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-utilities\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: E1205 12:08:32.973930 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.473918379 +0000 UTC m=+142.967781648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.974278 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-catalog-content\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:32 crc kubenswrapper[4807]: I1205 12:08:32.999746 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hzqvk" podStartSLOduration=7.999724785 podStartE2EDuration="7.999724785s" podCreationTimestamp="2025-12-05 12:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:32.96834493 +0000 UTC m=+142.462208199" watchObservedRunningTime="2025-12-05 12:08:32.999724785 +0000 UTC m=+142.493588054" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.016075 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x7xv\" (UniqueName: \"kubernetes.io/projected/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-kube-api-access-7x7xv\") pod \"community-operators-ff68x\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.039176 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44d68" podStartSLOduration=122.039157389 podStartE2EDuration="2m2.039157389s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.002353199 +0000 UTC m=+142.496216468" watchObservedRunningTime="2025-12-05 12:08:33.039157389 +0000 UTC m=+142.533020648" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.043409 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.059611 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.066664 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.073091 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dnrjc" podStartSLOduration=122.073070456 podStartE2EDuration="2m2.073070456s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.07146324 +0000 UTC m=+142.565326509" watchObservedRunningTime="2025-12-05 12:08:33.073070456 +0000 UTC m=+142.566933725" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.074263 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.074331 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.574314952 +0000 UTC m=+143.068178231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.078177 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-catalog-content\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.078243 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.078342 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljl7l\" (UniqueName: \"kubernetes.io/projected/792cf8c2-f4b4-4956-955d-43541a2adcd3-kube-api-access-ljl7l\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.078374 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-utilities\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.078920 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.578902952 +0000 UTC m=+143.072766291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.079208 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-catalog-content\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.079453 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-utilities\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.113885 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" podStartSLOduration=122.113864969 podStartE2EDuration="2m2.113864969s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.097781521 +0000 UTC m=+142.591644820" watchObservedRunningTime="2025-12-05 12:08:33.113864969 +0000 UTC m=+142.607728238" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.132415 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljl7l\" (UniqueName: \"kubernetes.io/projected/792cf8c2-f4b4-4956-955d-43541a2adcd3-kube-api-access-ljl7l\") pod \"certified-operators-sb9pm\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.133195 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" podStartSLOduration=122.13317908 podStartE2EDuration="2m2.13317908s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.133092717 +0000 UTC m=+142.626955976" watchObservedRunningTime="2025-12-05 12:08:33.13317908 +0000 UTC m=+142.627042349" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.181268 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.181748 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.681734064 +0000 UTC m=+143.175597333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.206113 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xlwdb" event={"ID":"8edc933b-ca85-4e1d-be88-5e619ea40286","Type":"ContainerStarted","Data":"bbda5d28dce27d92ea1b8391a4c70cbba6eb5a4d57a2f187f50581ce57830da8"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.255090 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.292478 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b2jjx" podStartSLOduration=122.292455201 podStartE2EDuration="2m2.292455201s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.21804424 +0000 UTC m=+142.711907509" watchObservedRunningTime="2025-12-05 12:08:33.292455201 +0000 UTC m=+142.786318470" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.295367 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.296283 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.796248569 +0000 UTC m=+143.290111838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.302657 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:33 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:33 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:33 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.302737 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.314645 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" podStartSLOduration=122.314624942 podStartE2EDuration="2m2.314624942s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.291581826 +0000 UTC m=+142.785445105" watchObservedRunningTime="2025-12-05 12:08:33.314624942 +0000 UTC m=+142.808488221" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.353982 4807 patch_prober.go:28] interesting pod/console-operator-58897d9998-zlzkl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.378621 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" podUID="a35ba616-01d1-4373-bbe9-84785afa9323" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.383837 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.411573 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" event={"ID":"a35ba616-01d1-4373-bbe9-84785afa9323","Type":"ContainerStarted","Data":"bf2062e72a75453f38bba45fc4bb251cb112bb3d7811c9d46c0a2121a3fae876"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.411611 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" event={"ID":"c64561f9-bf38-4a52-be8f-869658009f28","Type":"ContainerStarted","Data":"575b7e30b62edd8368a1ca7ce9559e02dac51fbdb61f3abbe931b91f39cea123"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.411625 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" event={"ID":"c64561f9-bf38-4a52-be8f-869658009f28","Type":"ContainerStarted","Data":"f1c2a2d6c5b866b9bb8013103cbb04e72fb35011f106ad89a545aad700344dca"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.398821 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.398880 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.898864405 +0000 UTC m=+143.392727674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.411965 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.424021 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:33.924002592 +0000 UTC m=+143.417865851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.429838 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" event={"ID":"bfbb6060-35b4-481e-b5c8-81f6a2280af0","Type":"ContainerStarted","Data":"1d2856ee43b34583c14bc847e401cd71e3983f17145472f8cc77594dda2264a9"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.430520 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m7fq9" podStartSLOduration=122.430509967 podStartE2EDuration="2m2.430509967s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.343923128 +0000 UTC m=+142.837786397" watchObservedRunningTime="2025-12-05 12:08:33.430509967 +0000 UTC m=+142.924373236" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.432820 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ftv59" podStartSLOduration=122.432813443 podStartE2EDuration="2m2.432813443s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.43131773 +0000 UTC m=+142.925181019" watchObservedRunningTime="2025-12-05 12:08:33.432813443 +0000 UTC m=+142.926676702" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.499899 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" event={"ID":"3e69d864-82d7-4f37-8f11-a7537640e94a","Type":"ContainerStarted","Data":"d363d63a156865c970581e98410f60d2516feb11056a486e378da00819622a59"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.512692 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.513012 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.012980988 +0000 UTC m=+143.506844257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.513269 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.513752 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.01374143 +0000 UTC m=+143.507604699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.515156 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" event={"ID":"d3d6a1f9-5ac9-40c2-862c-0b879eb91bee","Type":"ContainerStarted","Data":"e0807b4567b333ea8135d94ac96666d2bb55d6f6c38caa748e90dbd9e1a9e830"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.552609 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" podStartSLOduration=122.552590757 podStartE2EDuration="2m2.552590757s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.552402792 +0000 UTC m=+143.046266061" watchObservedRunningTime="2025-12-05 12:08:33.552590757 +0000 UTC m=+143.046454026" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.569335 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" event={"ID":"1c48a5de-7d23-4f1e-bdee-31d122fd26ad","Type":"ContainerStarted","Data":"a06184c30f870ce3ba15bcb0c0f620f924aadc4f5957acbaeda4dd6e3356e4bc"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.603458 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mz7vc" podStartSLOduration=122.603438807 podStartE2EDuration="2m2.603438807s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:33.598800654 +0000 UTC m=+143.092663943" watchObservedRunningTime="2025-12-05 12:08:33.603438807 +0000 UTC m=+143.097302076" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.615970 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.617371 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.117351884 +0000 UTC m=+143.611215153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.624067 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" event={"ID":"c63ee192-3315-4176-9654-8497d5ba9fd7","Type":"ContainerStarted","Data":"7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.624492 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.636765 4807 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-x4t9b container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" start-of-body= Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.640217 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" podUID="c63ee192-3315-4176-9654-8497d5ba9fd7" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.646275 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-56vtp" event={"ID":"cecd7f96-25b6-4f5e-9a93-c00cf902e00a","Type":"ContainerStarted","Data":"0641cc99ffa82b1a508fa51e39520446087f99aa4ff2b39e220e0c76f091e790"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.717907 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" event={"ID":"ec820b06-dcc4-4180-b22d-019b6696a3b4","Type":"ContainerStarted","Data":"6b550e434c08fa4ac993141db7fc8aabfa0bd41dee7c398b6e263a6eaaa4fd57"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.717955 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" event={"ID":"ec820b06-dcc4-4180-b22d-019b6696a3b4","Type":"ContainerStarted","Data":"96dec3c0f0a03aca7ec3898fe93d609209223d2c73abc8648e3dab6c7d18f7a3"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.718636 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.725156 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.225139817 +0000 UTC m=+143.719003086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.755433 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" event={"ID":"dc3e80de-9521-4aab-9210-0e4e2db9b002","Type":"ContainerStarted","Data":"ddc6b139ed2c7b882eab8c7dc1029d8b10e14de1c05fd0417e2cde470e76d44e"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.755925 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.803054 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" event={"ID":"156bdc3e-1f65-44cd-8ccd-26bc625a01c5","Type":"ContainerStarted","Data":"26ed8c3bc1a2a728b3904118a0330f902db0fc2ef59e639b251ea71830e45c73"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.820322 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.821292 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.321275727 +0000 UTC m=+143.815138996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.850037 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" event={"ID":"ed25959e-c671-4690-a9c7-f2b2a8b67d0b","Type":"ContainerStarted","Data":"6fed8f5aa669e34e75bea15489f6acf7305be0147391548f297c3b5772907c74"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.850085 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" event={"ID":"ed25959e-c671-4690-a9c7-f2b2a8b67d0b","Type":"ContainerStarted","Data":"43e32604290416e165734cdb0c9c08d1848b373aa96c78f6f9d8b1d2e4092526"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.851625 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" event={"ID":"5522a41b-50f2-4ab2-885c-0161017a9ea9","Type":"ContainerStarted","Data":"bfe751e55577ebeae48ee3780005819b19380a31ef099bc483f7ef1af722a0a1"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.852926 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kfhwr" event={"ID":"753e11b0-5a03-432d-932a-87031ae782ea","Type":"ContainerStarted","Data":"035dbd6e53c38bce9f6a7c37ff911ccc967e4e04a11f226aa1218a5d562a3070"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.854059 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" event={"ID":"26646615-02ee-4175-9af9-01403073049b","Type":"ContainerStarted","Data":"7de580f878e7f983ce9152877e2de7d32342c831cec9dddaafbc0c91d37d8911"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.854762 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.855501 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hzqvk" event={"ID":"afbcd69b-9f75-4999-af80-31052865b708","Type":"ContainerStarted","Data":"924331eba838ff996748440ef62f793d26dc74bfd7b314a5fc4a26939e86e1c0"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.857493 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" event={"ID":"e849f7a7-467d-4124-a342-c8b842e28790","Type":"ContainerStarted","Data":"c3ad01668fff5546c385b6313f72931ffeac99b301c41e4f2a52029b1f138d40"} Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.861965 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.872315 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5j622" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.872555 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.890858 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.894350 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-927f9"] Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.906908 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4d4jf"] Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.911247 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-njg49" Dec 05 12:08:33 crc kubenswrapper[4807]: I1205 12:08:33.922598 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:33 crc kubenswrapper[4807]: E1205 12:08:33.925790 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.425766537 +0000 UTC m=+143.919629806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.011915 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.041224 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.041628 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.54161102 +0000 UTC m=+144.035474289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.093617 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ff68x"] Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.144958 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.145384 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.645369538 +0000 UTC m=+144.139232817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: W1205 12:08:34.150812 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02e14a82_3ab0_41f7_9ad4_5ca9ef73430d.slice/crio-67d1663936ed8e7a6751788f559cffe8016c3046e58ce693ee0c434891d06c36 WatchSource:0}: Error finding container 67d1663936ed8e7a6751788f559cffe8016c3046e58ce693ee0c434891d06c36: Status 404 returned error can't find the container with id 67d1663936ed8e7a6751788f559cffe8016c3046e58ce693ee0c434891d06c36 Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.236342 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sb9pm"] Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.249019 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.249582 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.749560148 +0000 UTC m=+144.243423407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.302766 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:34 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:34 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:34 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.302821 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.353142 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.353446 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.853435579 +0000 UTC m=+144.347298848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.406644 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-m9vtg" podStartSLOduration=123.406627756 podStartE2EDuration="2m3.406627756s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.406265205 +0000 UTC m=+143.900128474" watchObservedRunningTime="2025-12-05 12:08:34.406627756 +0000 UTC m=+143.900491025" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.408246 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gzz4h" podStartSLOduration=123.408237412 podStartE2EDuration="2m3.408237412s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.330728342 +0000 UTC m=+143.824591611" watchObservedRunningTime="2025-12-05 12:08:34.408237412 +0000 UTC m=+143.902100681" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.453717 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.453848 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.953821401 +0000 UTC m=+144.447684670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.453985 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.454258 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:34.954238283 +0000 UTC m=+144.448101552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.491092 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" podStartSLOduration=124.491075724 podStartE2EDuration="2m4.491075724s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.444017692 +0000 UTC m=+143.937880961" watchObservedRunningTime="2025-12-05 12:08:34.491075724 +0000 UTC m=+143.984938993" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.501779 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xsv4c"] Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.502790 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.505965 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.530468 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsv4c"] Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.555087 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.555405 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-utilities\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.555519 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-catalog-content\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.555579 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm6jv\" (UniqueName: \"kubernetes.io/projected/2b219446-dca8-4aec-b300-92eb952d2a89-kube-api-access-lm6jv\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.555710 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.055694776 +0000 UTC m=+144.549558035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.572635 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8269c" podStartSLOduration=124.572616508 podStartE2EDuration="2m4.572616508s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.559585967 +0000 UTC m=+144.053449256" watchObservedRunningTime="2025-12-05 12:08:34.572616508 +0000 UTC m=+144.066479767" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.609028 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z274j" podStartSLOduration=123.609015556 podStartE2EDuration="2m3.609015556s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.606606897 +0000 UTC m=+144.100470166" watchObservedRunningTime="2025-12-05 12:08:34.609015556 +0000 UTC m=+144.102878825" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.656850 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-catalog-content\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.656891 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.656922 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm6jv\" (UniqueName: \"kubernetes.io/projected/2b219446-dca8-4aec-b300-92eb952d2a89-kube-api-access-lm6jv\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.656973 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-utilities\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.657235 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.15721882 +0000 UTC m=+144.651082089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.657473 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-catalog-content\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.657519 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-utilities\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.709545 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm6jv\" (UniqueName: \"kubernetes.io/projected/2b219446-dca8-4aec-b300-92eb952d2a89-kube-api-access-lm6jv\") pod \"redhat-marketplace-xsv4c\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.759162 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.759774 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.259754684 +0000 UTC m=+144.753617963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.763757 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.782225 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" podStartSLOduration=124.782209014 podStartE2EDuration="2m4.782209014s" podCreationTimestamp="2025-12-05 12:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.780604248 +0000 UTC m=+144.274467547" watchObservedRunningTime="2025-12-05 12:08:34.782209014 +0000 UTC m=+144.276072283" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.822351 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" podStartSLOduration=123.822329767 podStartE2EDuration="2m3.822329767s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.814459053 +0000 UTC m=+144.308322332" watchObservedRunningTime="2025-12-05 12:08:34.822329767 +0000 UTC m=+144.316193036" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.861474 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.862908 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.362891604 +0000 UTC m=+144.856754873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.863931 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6mxzg" podStartSLOduration=123.863909323 podStartE2EDuration="2m3.863909323s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.863557253 +0000 UTC m=+144.357420522" watchObservedRunningTime="2025-12-05 12:08:34.863909323 +0000 UTC m=+144.357772602" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.899369 4807 generic.go:334] "Generic (PLEG): container finished" podID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerID="df498dc0e65691752278b2aee36eb3c68a49e7a6a273a8780fc32a07461f6cd6" exitCode=0 Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.899453 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-927f9" event={"ID":"55077615-de5e-48c6-9be0-2f3c8bbabae7","Type":"ContainerDied","Data":"df498dc0e65691752278b2aee36eb3c68a49e7a6a273a8780fc32a07461f6cd6"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.899485 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-927f9" event={"ID":"55077615-de5e-48c6-9be0-2f3c8bbabae7","Type":"ContainerStarted","Data":"5204b0542fac4d880f8a962429d4e26e08dc5e21f10268badb9ecb55e1d8b905"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.913060 4807 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.920577 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-77zhj"] Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.921788 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.922580 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb9pm" event={"ID":"792cf8c2-f4b4-4956-955d-43541a2adcd3","Type":"ContainerStarted","Data":"795e5823cf490ed1f3e1137177461da5d8dbaaba90087d6d51f86e275eb97101"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.929688 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zp9xg" podStartSLOduration=123.929670708 podStartE2EDuration="2m3.929670708s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:34.927885327 +0000 UTC m=+144.421748606" watchObservedRunningTime="2025-12-05 12:08:34.929670708 +0000 UTC m=+144.423533977" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.937968 4807 generic.go:334] "Generic (PLEG): container finished" podID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerID="182f2ef6a5d2e4bc76f7fa70c6c0c3363e831452e7f92540dc5e054341813fa6" exitCode=0 Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.938042 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff68x" event={"ID":"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d","Type":"ContainerDied","Data":"182f2ef6a5d2e4bc76f7fa70c6c0c3363e831452e7f92540dc5e054341813fa6"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.938069 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff68x" event={"ID":"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d","Type":"ContainerStarted","Data":"67d1663936ed8e7a6751788f559cffe8016c3046e58ce693ee0c434891d06c36"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.951354 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xlwdb" event={"ID":"8edc933b-ca85-4e1d-be88-5e619ea40286","Type":"ContainerStarted","Data":"8bce07e7d9f08535220fa42317879ff410e949162a735038caa2d5b56eadf9c1"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.952191 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.953757 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" event={"ID":"bfbb6060-35b4-481e-b5c8-81f6a2280af0","Type":"ContainerStarted","Data":"06e21c6ab09513422d367039b49a6fd5ade6771baaeda8f598c3cdfccda5aa9f"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.955238 4807 generic.go:334] "Generic (PLEG): container finished" podID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerID="11d02f984ff2b89023f221c6c2e2289932914bd31466232fd48a8d8ee7ead036" exitCode=0 Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.992283 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.992872 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-catalog-content\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.992961 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rcfp\" (UniqueName: \"kubernetes.io/projected/58c70b2e-9b4d-4820-b089-aec92a30ca58-kube-api-access-7rcfp\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.992994 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-utilities\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.993487 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d4jf" event={"ID":"c9af52ad-e3d9-426f-9adf-a9880437d83f","Type":"ContainerDied","Data":"11d02f984ff2b89023f221c6c2e2289932914bd31466232fd48a8d8ee7ead036"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.993551 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d4jf" event={"ID":"c9af52ad-e3d9-426f-9adf-a9880437d83f","Type":"ContainerStarted","Data":"89b2e81423724508e8588cb7b9a6816cf33c9fcdf80b64ea6b7f6a97a45fa081"} Dec 05 12:08:34 crc kubenswrapper[4807]: I1205 12:08:34.993574 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-77zhj"] Dec 05 12:08:34 crc kubenswrapper[4807]: E1205 12:08:34.994392 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.494373162 +0000 UTC m=+144.988236431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.094019 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.094731 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-catalog-content\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.094973 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rcfp\" (UniqueName: \"kubernetes.io/projected/58c70b2e-9b4d-4820-b089-aec92a30ca58-kube-api-access-7rcfp\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.095048 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-utilities\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.107392 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.607374164 +0000 UTC m=+145.101237433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.120106 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-utilities\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.130871 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-catalog-content\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.196383 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rcfp\" (UniqueName: \"kubernetes.io/projected/58c70b2e-9b4d-4820-b089-aec92a30ca58-kube-api-access-7rcfp\") pod \"redhat-marketplace-77zhj\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.209135 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.209642 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.709624669 +0000 UTC m=+145.203487938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.227720 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-xlwdb" podStartSLOduration=10.227702105 podStartE2EDuration="10.227702105s" podCreationTimestamp="2025-12-05 12:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:35.1858041 +0000 UTC m=+144.679667369" watchObservedRunningTime="2025-12-05 12:08:35.227702105 +0000 UTC m=+144.721565374" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.247866 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.285256 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsv4c"] Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.294964 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:35 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:35 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:35 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.295021 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.318646 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.318985 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.818971347 +0000 UTC m=+145.312834616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.419678 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.420422 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:35.920401568 +0000 UTC m=+145.414264837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.516244 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-44xrt" podStartSLOduration=124.51621681 podStartE2EDuration="2m4.51621681s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:35.287779838 +0000 UTC m=+144.781643107" watchObservedRunningTime="2025-12-05 12:08:35.51621681 +0000 UTC m=+145.010080079" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.516618 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9c8wq"] Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.517834 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.519407 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-zlzkl" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.522057 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.522311 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.022300583 +0000 UTC m=+145.516163852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.523890 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.583923 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9c8wq"] Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.626096 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.626389 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-catalog-content\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.626455 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-utilities\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.626497 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5z6f\" (UniqueName: \"kubernetes.io/projected/572bedd6-2920-481b-a356-c96fa8766456-kube-api-access-j5z6f\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.626631 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.126614008 +0000 UTC m=+145.620477277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.728278 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.728330 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5z6f\" (UniqueName: \"kubernetes.io/projected/572bedd6-2920-481b-a356-c96fa8766456-kube-api-access-j5z6f\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.728408 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-catalog-content\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.728479 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-utilities\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.728950 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-utilities\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.729214 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.229200742 +0000 UTC m=+145.723064011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.729730 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-catalog-content\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.789296 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5z6f\" (UniqueName: \"kubernetes.io/projected/572bedd6-2920-481b-a356-c96fa8766456-kube-api-access-j5z6f\") pod \"redhat-operators-9c8wq\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.830997 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.831408 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.331390996 +0000 UTC m=+145.825254265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.831812 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.832112 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.332101956 +0000 UTC m=+145.825965225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.858634 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.925149 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jjbd7"] Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.937115 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:35 crc kubenswrapper[4807]: E1205 12:08:35.937692 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.437671956 +0000 UTC m=+145.931535225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.937825 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:35 crc kubenswrapper[4807]: I1205 12:08:35.978459 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jjbd7"] Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.000640 4807 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-x4t9b container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.38:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.000706 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" podUID="c63ee192-3315-4176-9654-8497d5ba9fd7" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.38:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.038683 4807 generic.go:334] "Generic (PLEG): container finished" podID="2b219446-dca8-4aec-b300-92eb952d2a89" containerID="60cf6d9ebeb17ca53a6db4cc25720ffc7dc81ba5f78a2be8790a4598d6a158a7" exitCode=0 Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.038813 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsv4c" event={"ID":"2b219446-dca8-4aec-b300-92eb952d2a89","Type":"ContainerDied","Data":"60cf6d9ebeb17ca53a6db4cc25720ffc7dc81ba5f78a2be8790a4598d6a158a7"} Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.038847 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsv4c" event={"ID":"2b219446-dca8-4aec-b300-92eb952d2a89","Type":"ContainerStarted","Data":"62ebfc69a567e5f67fbb0a522668836e339b37319b4964363fa0d649284b5efc"} Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.042132 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-catalog-content\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.042209 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-utilities\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.042251 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.042272 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gvks\" (UniqueName: \"kubernetes.io/projected/5f725110-9ece-41b3-a2af-2fa60bf2f649-kube-api-access-9gvks\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.046077 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.546062836 +0000 UTC m=+146.039926105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.055349 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" event={"ID":"1c48a5de-7d23-4f1e-bdee-31d122fd26ad","Type":"ContainerStarted","Data":"d2feb09c917f7743253c8f9f1e09e30d462b5ea8fc83662b8d0f38597d71ef68"} Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.101833 4807 generic.go:334] "Generic (PLEG): container finished" podID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerID="da7366947a208d15882969e83f229edeb5b4386c2e52bb09e87424910646bdb9" exitCode=0 Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.103468 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb9pm" event={"ID":"792cf8c2-f4b4-4956-955d-43541a2adcd3","Type":"ContainerDied","Data":"da7366947a208d15882969e83f229edeb5b4386c2e52bb09e87424910646bdb9"} Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.143782 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.144217 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-utilities\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.144296 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gvks\" (UniqueName: \"kubernetes.io/projected/5f725110-9ece-41b3-a2af-2fa60bf2f649-kube-api-access-9gvks\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.144626 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-catalog-content\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.145503 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.64548335 +0000 UTC m=+146.139346619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.147492 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-utilities\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.147974 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-catalog-content\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.192418 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gvks\" (UniqueName: \"kubernetes.io/projected/5f725110-9ece-41b3-a2af-2fa60bf2f649-kube-api-access-9gvks\") pod \"redhat-operators-jjbd7\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.211475 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-77zhj"] Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.245819 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.246137 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.74612583 +0000 UTC m=+146.239989099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.317104 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:36 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:36 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:36 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.317330 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.317347 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.348731 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.349073 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.849057595 +0000 UTC m=+146.342920864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.419458 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.426204 4807 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.450367 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.450818 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:36.950804375 +0000 UTC m=+146.444667644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.560167 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.560348 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.060321988 +0000 UTC m=+146.554185257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.560443 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.560782 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.060771501 +0000 UTC m=+146.554634770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.599182 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9c8wq"] Dec 05 12:08:36 crc kubenswrapper[4807]: W1205 12:08:36.651393 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod572bedd6_2920_481b_a356_c96fa8766456.slice/crio-221acccf132fea0e0ef368c20532cfaced3dcb9d04b9ef67bf58d4a8bcd4fa79 WatchSource:0}: Error finding container 221acccf132fea0e0ef368c20532cfaced3dcb9d04b9ef67bf58d4a8bcd4fa79: Status 404 returned error can't find the container with id 221acccf132fea0e0ef368c20532cfaced3dcb9d04b9ef67bf58d4a8bcd4fa79 Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.661325 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.662088 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.162067838 +0000 UTC m=+146.655931107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.768706 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.769634 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.269618725 +0000 UTC m=+146.763481994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p67q6" (UID: "2ea71070-7150-4f2d-b531-08f22697c479") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.870444 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:36 crc kubenswrapper[4807]: E1205 12:08:36.870964 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 12:08:37.370949793 +0000 UTC m=+146.864813052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.917350 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jjbd7"] Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.921359 4807 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-05T12:08:36.426232835Z","Handler":null,"Name":""} Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.965506 4807 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.965555 4807 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 05 12:08:36 crc kubenswrapper[4807]: I1205 12:08:36.972680 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.002060 4807 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.002100 4807 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.089143 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p67q6\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.131965 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9c8wq" event={"ID":"572bedd6-2920-481b-a356-c96fa8766456","Type":"ContainerStarted","Data":"221acccf132fea0e0ef368c20532cfaced3dcb9d04b9ef67bf58d4a8bcd4fa79"} Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.135256 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.135294 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.163357 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jjbd7" event={"ID":"5f725110-9ece-41b3-a2af-2fa60bf2f649","Type":"ContainerStarted","Data":"62f0ddbc29a2fcc498edc8bebb8f8c61d3a20fefc1936f9d86cddfefde7e736e"} Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.165751 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.175541 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.193677 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.221894 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.246595 4807 generic.go:334] "Generic (PLEG): container finished" podID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerID="887faa743a261a6ce3c251075d938da2e81b467387a3e7228051b7a076c1fbc6" exitCode=0 Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.250486 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.251482 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" event={"ID":"1c48a5de-7d23-4f1e-bdee-31d122fd26ad","Type":"ContainerStarted","Data":"abb31b077fca9cdc768cb7002c8a23852ca6d91ebd6009c14f5c779bdbf7a341"} Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.251504 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77zhj" event={"ID":"58c70b2e-9b4d-4820-b089-aec92a30ca58","Type":"ContainerDied","Data":"887faa743a261a6ce3c251075d938da2e81b467387a3e7228051b7a076c1fbc6"} Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.251517 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77zhj" event={"ID":"58c70b2e-9b4d-4820-b089-aec92a30ca58","Type":"ContainerStarted","Data":"4636e9a1afbcd535191f9294fa6e0557f9cbb14ddb32d6a3a6415cf3cbb866ce"} Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.295283 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:37 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:37 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:37 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.295341 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.302972 4807 patch_prober.go:28] interesting pod/downloads-7954f5f757-25wjt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.303004 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25wjt" podUID="b0241b9d-883d-4e59-8f4c-38bf4f8efc91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.305919 4807 patch_prober.go:28] interesting pod/downloads-7954f5f757-25wjt container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.305949 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-25wjt" podUID="b0241b9d-883d-4e59-8f4c-38bf4f8efc91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.653070 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p67q6"] Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.788177 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.799326 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.889154 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.889321 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.889408 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.890772 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.909813 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.910019 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.953854 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.962849 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 12:08:37 crc kubenswrapper[4807]: I1205 12:08:37.968868 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.005368 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.005558 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.014309 4807 patch_prober.go:28] interesting pod/console-f9d7485db-kf4mv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.014371 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kf4mv" podUID="7bbff2d6-47e4-489b-9c99-b3ef91531447" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.287608 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.290563 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:38 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:38 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:38 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.290613 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.307231 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" event={"ID":"1c48a5de-7d23-4f1e-bdee-31d122fd26ad","Type":"ContainerStarted","Data":"5a27a1333cba2e8671e14e527a7c318ef79d1183e7790ed0ebcdbaf3a30e6a8e"} Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.310869 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" event={"ID":"2ea71070-7150-4f2d-b531-08f22697c479","Type":"ContainerStarted","Data":"4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400"} Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.310904 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" event={"ID":"2ea71070-7150-4f2d-b531-08f22697c479","Type":"ContainerStarted","Data":"c092fcb75d29fd93436d1cf2531daa73c2d913de0c78d7b5f3123efc7581c3c8"} Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.311221 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.327217 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2cc88" podStartSLOduration=13.327177021 podStartE2EDuration="13.327177021s" podCreationTimestamp="2025-12-05 12:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:38.326219874 +0000 UTC m=+147.820083163" watchObservedRunningTime="2025-12-05 12:08:38.327177021 +0000 UTC m=+147.821040310" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.342508 4807 generic.go:334] "Generic (PLEG): container finished" podID="572bedd6-2920-481b-a356-c96fa8766456" containerID="edea6372167268d889da61726b7c132644f31108c85bb3ffcfdff2012e895c74" exitCode=0 Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.343308 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9c8wq" event={"ID":"572bedd6-2920-481b-a356-c96fa8766456","Type":"ContainerDied","Data":"edea6372167268d889da61726b7c132644f31108c85bb3ffcfdff2012e895c74"} Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.354674 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" podStartSLOduration=127.354656084 podStartE2EDuration="2m7.354656084s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:38.35346297 +0000 UTC m=+147.847326249" watchObservedRunningTime="2025-12-05 12:08:38.354656084 +0000 UTC m=+147.848519353" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.366229 4807 generic.go:334] "Generic (PLEG): container finished" podID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerID="df330a45426c14b0f6fe534e5471637d7fc15f910f47ad581c7464126eb5a658" exitCode=0 Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.368052 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jjbd7" event={"ID":"5f725110-9ece-41b3-a2af-2fa60bf2f649","Type":"ContainerDied","Data":"df330a45426c14b0f6fe534e5471637d7fc15f910f47ad581c7464126eb5a658"} Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.385762 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-g62qt" Dec 05 12:08:38 crc kubenswrapper[4807]: W1205 12:08:38.409207 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-08ee764b1186fd35ab03d72d9cdb448af65b1d0e2b3ecd7d32ae96e1053904d3 WatchSource:0}: Error finding container 08ee764b1186fd35ab03d72d9cdb448af65b1d0e2b3ecd7d32ae96e1053904d3: Status 404 returned error can't find the container with id 08ee764b1186fd35ab03d72d9cdb448af65b1d0e2b3ecd7d32ae96e1053904d3 Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.525613 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.526666 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.528837 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.529128 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.558509 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.712092 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.712214 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.813515 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.813666 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.813761 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.831186 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:38 crc kubenswrapper[4807]: I1205 12:08:38.896955 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.302725 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.321791 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:39 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:39 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:39 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.321842 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.388918 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"28a75ead95f32b7a2947245cc7edd5bc00be4dccdc4c5cc4bf7e66c96095f3b2"} Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.388964 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"08ee764b1186fd35ab03d72d9cdb448af65b1d0e2b3ecd7d32ae96e1053904d3"} Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.436277 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cada4dac053d09124f7824fe92bcfcc0e0af47758d37ccb33e3b9dcfb53f84d0"} Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.436346 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"284a98567c95b02bc5b493f97b043c06eed7ef2aa3171deefb8a6b4d41fcd165"} Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.436554 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.464405 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"45ba927c06ebed09c57e95a9e5a8525fdba6f5e5f0dc2515ddf26df4af7a01ad"} Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.464472 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a0eae33f997e6b14f325920d95e320b462fd9d8f30bb4248993b4a868b6f5b96"} Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.466641 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3583f6c3-d1fd-4306-bda3-a0da3f00cb97","Type":"ContainerStarted","Data":"c59656006a8d7554f0793bcd52b71db9007611285a6ad6b673aa36ed17260e7d"} Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.478408 4807 generic.go:334] "Generic (PLEG): container finished" podID="e849f7a7-467d-4124-a342-c8b842e28790" containerID="c3ad01668fff5546c385b6313f72931ffeac99b301c41e4f2a52029b1f138d40" exitCode=0 Dec 05 12:08:39 crc kubenswrapper[4807]: I1205 12:08:39.478480 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" event={"ID":"e849f7a7-467d-4124-a342-c8b842e28790","Type":"ContainerDied","Data":"c3ad01668fff5546c385b6313f72931ffeac99b301c41e4f2a52029b1f138d40"} Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.292505 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:40 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:40 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:40 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.292843 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.827318 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.958619 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e849f7a7-467d-4124-a342-c8b842e28790-config-volume\") pod \"e849f7a7-467d-4124-a342-c8b842e28790\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.958715 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c77ck\" (UniqueName: \"kubernetes.io/projected/e849f7a7-467d-4124-a342-c8b842e28790-kube-api-access-c77ck\") pod \"e849f7a7-467d-4124-a342-c8b842e28790\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.958746 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e849f7a7-467d-4124-a342-c8b842e28790-secret-volume\") pod \"e849f7a7-467d-4124-a342-c8b842e28790\" (UID: \"e849f7a7-467d-4124-a342-c8b842e28790\") " Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.960431 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e849f7a7-467d-4124-a342-c8b842e28790-config-volume" (OuterVolumeSpecName: "config-volume") pod "e849f7a7-467d-4124-a342-c8b842e28790" (UID: "e849f7a7-467d-4124-a342-c8b842e28790"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.967212 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e849f7a7-467d-4124-a342-c8b842e28790-kube-api-access-c77ck" (OuterVolumeSpecName: "kube-api-access-c77ck") pod "e849f7a7-467d-4124-a342-c8b842e28790" (UID: "e849f7a7-467d-4124-a342-c8b842e28790"). InnerVolumeSpecName "kube-api-access-c77ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:08:40 crc kubenswrapper[4807]: I1205 12:08:40.967706 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e849f7a7-467d-4124-a342-c8b842e28790-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e849f7a7-467d-4124-a342-c8b842e28790" (UID: "e849f7a7-467d-4124-a342-c8b842e28790"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.060674 4807 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e849f7a7-467d-4124-a342-c8b842e28790-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.060708 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c77ck\" (UniqueName: \"kubernetes.io/projected/e849f7a7-467d-4124-a342-c8b842e28790-kube-api-access-c77ck\") on node \"crc\" DevicePath \"\"" Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.060723 4807 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e849f7a7-467d-4124-a342-c8b842e28790-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.290427 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:41 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:41 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:41 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.290681 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.534229 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3583f6c3-d1fd-4306-bda3-a0da3f00cb97","Type":"ContainerStarted","Data":"0b1abc26adc4f1ed7a73c67ae5d1cc469a69ee08dc3ba58bffd12d8f923c8ef4"} Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.550372 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" event={"ID":"e849f7a7-467d-4124-a342-c8b842e28790","Type":"ContainerDied","Data":"732ee5c2c7abbe8742e8497bd91870d87398bf1b0adc87585b2ae5f7d2f6d0cd"} Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.550438 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="732ee5c2c7abbe8742e8497bd91870d87398bf1b0adc87585b2ae5f7d2f6d0cd" Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.550577 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx" Dec 05 12:08:41 crc kubenswrapper[4807]: I1205 12:08:41.561886 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.561869192 podStartE2EDuration="3.561869192s" podCreationTimestamp="2025-12-05 12:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:08:41.557704774 +0000 UTC m=+151.051568043" watchObservedRunningTime="2025-12-05 12:08:41.561869192 +0000 UTC m=+151.055732461" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.288677 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:42 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:42 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:42 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.288757 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.578929 4807 generic.go:334] "Generic (PLEG): container finished" podID="3583f6c3-d1fd-4306-bda3-a0da3f00cb97" containerID="0b1abc26adc4f1ed7a73c67ae5d1cc469a69ee08dc3ba58bffd12d8f923c8ef4" exitCode=0 Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.578971 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3583f6c3-d1fd-4306-bda3-a0da3f00cb97","Type":"ContainerDied","Data":"0b1abc26adc4f1ed7a73c67ae5d1cc469a69ee08dc3ba58bffd12d8f923c8ef4"} Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.628270 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 12:08:42 crc kubenswrapper[4807]: E1205 12:08:42.628602 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e849f7a7-467d-4124-a342-c8b842e28790" containerName="collect-profiles" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.628615 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="e849f7a7-467d-4124-a342-c8b842e28790" containerName="collect-profiles" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.628714 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="e849f7a7-467d-4124-a342-c8b842e28790" containerName="collect-profiles" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.629131 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.631217 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.632742 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.633645 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.792833 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f874947-e0c1-4e68-9c63-11016549a7db-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.793233 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f874947-e0c1-4e68-9c63-11016549a7db-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.895670 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f874947-e0c1-4e68-9c63-11016549a7db-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.895769 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f874947-e0c1-4e68-9c63-11016549a7db-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:42 crc kubenswrapper[4807]: I1205 12:08:42.896052 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f874947-e0c1-4e68-9c63-11016549a7db-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:43 crc kubenswrapper[4807]: I1205 12:08:43.138555 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f874947-e0c1-4e68-9c63-11016549a7db-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:43 crc kubenswrapper[4807]: I1205 12:08:43.248573 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:43 crc kubenswrapper[4807]: I1205 12:08:43.289572 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:43 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:43 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:43 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:43 crc kubenswrapper[4807]: I1205 12:08:43.289678 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:43 crc kubenswrapper[4807]: I1205 12:08:43.470493 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-xlwdb" Dec 05 12:08:43 crc kubenswrapper[4807]: I1205 12:08:43.817233 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 12:08:43 crc kubenswrapper[4807]: W1205 12:08:43.868158 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1f874947_e0c1_4e68_9c63_11016549a7db.slice/crio-1504002a2cfcfeaf90a24bb64fc62fc1c896d05f6713d4cb608bb46421cbe0c2 WatchSource:0}: Error finding container 1504002a2cfcfeaf90a24bb64fc62fc1c896d05f6713d4cb608bb46421cbe0c2: Status 404 returned error can't find the container with id 1504002a2cfcfeaf90a24bb64fc62fc1c896d05f6713d4cb608bb46421cbe0c2 Dec 05 12:08:43 crc kubenswrapper[4807]: I1205 12:08:43.967700 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.036442 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kubelet-dir\") pod \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.036609 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kube-api-access\") pod \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\" (UID: \"3583f6c3-d1fd-4306-bda3-a0da3f00cb97\") " Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.037569 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3583f6c3-d1fd-4306-bda3-a0da3f00cb97" (UID: "3583f6c3-d1fd-4306-bda3-a0da3f00cb97"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.049769 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3583f6c3-d1fd-4306-bda3-a0da3f00cb97" (UID: "3583f6c3-d1fd-4306-bda3-a0da3f00cb97"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.138146 4807 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.138183 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3583f6c3-d1fd-4306-bda3-a0da3f00cb97-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.289561 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:44 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:44 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:44 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.289608 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.626772 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.628038 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3583f6c3-d1fd-4306-bda3-a0da3f00cb97","Type":"ContainerDied","Data":"c59656006a8d7554f0793bcd52b71db9007611285a6ad6b673aa36ed17260e7d"} Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.628082 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c59656006a8d7554f0793bcd52b71db9007611285a6ad6b673aa36ed17260e7d" Dec 05 12:08:44 crc kubenswrapper[4807]: I1205 12:08:44.633395 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1f874947-e0c1-4e68-9c63-11016549a7db","Type":"ContainerStarted","Data":"1504002a2cfcfeaf90a24bb64fc62fc1c896d05f6713d4cb608bb46421cbe0c2"} Dec 05 12:08:45 crc kubenswrapper[4807]: I1205 12:08:45.288839 4807 patch_prober.go:28] interesting pod/router-default-5444994796-b52pb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 12:08:45 crc kubenswrapper[4807]: [-]has-synced failed: reason withheld Dec 05 12:08:45 crc kubenswrapper[4807]: [+]process-running ok Dec 05 12:08:45 crc kubenswrapper[4807]: healthz check failed Dec 05 12:08:45 crc kubenswrapper[4807]: I1205 12:08:45.288893 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b52pb" podUID="6d9d73ba-2eed-4b7d-925e-47bba9c1ae8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 12:08:45 crc kubenswrapper[4807]: I1205 12:08:45.651907 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1f874947-e0c1-4e68-9c63-11016549a7db","Type":"ContainerStarted","Data":"ae4e2da76d0207993a1992543f98965bb66e938fdcd601f1d958fdb06302525d"} Dec 05 12:08:46 crc kubenswrapper[4807]: I1205 12:08:46.289517 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:46 crc kubenswrapper[4807]: I1205 12:08:46.292021 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-b52pb" Dec 05 12:08:46 crc kubenswrapper[4807]: I1205 12:08:46.659299 4807 generic.go:334] "Generic (PLEG): container finished" podID="1f874947-e0c1-4e68-9c63-11016549a7db" containerID="ae4e2da76d0207993a1992543f98965bb66e938fdcd601f1d958fdb06302525d" exitCode=0 Dec 05 12:08:46 crc kubenswrapper[4807]: I1205 12:08:46.659342 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1f874947-e0c1-4e68-9c63-11016549a7db","Type":"ContainerDied","Data":"ae4e2da76d0207993a1992543f98965bb66e938fdcd601f1d958fdb06302525d"} Dec 05 12:08:47 crc kubenswrapper[4807]: I1205 12:08:47.303441 4807 patch_prober.go:28] interesting pod/downloads-7954f5f757-25wjt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 05 12:08:47 crc kubenswrapper[4807]: I1205 12:08:47.303501 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25wjt" podUID="b0241b9d-883d-4e59-8f4c-38bf4f8efc91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 05 12:08:47 crc kubenswrapper[4807]: I1205 12:08:47.303577 4807 patch_prober.go:28] interesting pod/downloads-7954f5f757-25wjt container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 05 12:08:47 crc kubenswrapper[4807]: I1205 12:08:47.303625 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-25wjt" podUID="b0241b9d-883d-4e59-8f4c-38bf4f8efc91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 05 12:08:48 crc kubenswrapper[4807]: I1205 12:08:48.005972 4807 patch_prober.go:28] interesting pod/console-f9d7485db-kf4mv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 05 12:08:48 crc kubenswrapper[4807]: I1205 12:08:48.006030 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kf4mv" podUID="7bbff2d6-47e4-489b-9c99-b3ef91531447" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 05 12:08:51 crc kubenswrapper[4807]: I1205 12:08:51.748709 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:51 crc kubenswrapper[4807]: I1205 12:08:51.854574 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f874947-e0c1-4e68-9c63-11016549a7db-kubelet-dir\") pod \"1f874947-e0c1-4e68-9c63-11016549a7db\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " Dec 05 12:08:51 crc kubenswrapper[4807]: I1205 12:08:51.854654 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f874947-e0c1-4e68-9c63-11016549a7db-kube-api-access\") pod \"1f874947-e0c1-4e68-9c63-11016549a7db\" (UID: \"1f874947-e0c1-4e68-9c63-11016549a7db\") " Dec 05 12:08:51 crc kubenswrapper[4807]: I1205 12:08:51.854686 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f874947-e0c1-4e68-9c63-11016549a7db-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1f874947-e0c1-4e68-9c63-11016549a7db" (UID: "1f874947-e0c1-4e68-9c63-11016549a7db"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:08:51 crc kubenswrapper[4807]: I1205 12:08:51.854928 4807 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f874947-e0c1-4e68-9c63-11016549a7db-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:08:51 crc kubenswrapper[4807]: I1205 12:08:51.860693 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f874947-e0c1-4e68-9c63-11016549a7db-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1f874947-e0c1-4e68-9c63-11016549a7db" (UID: "1f874947-e0c1-4e68-9c63-11016549a7db"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:08:51 crc kubenswrapper[4807]: I1205 12:08:51.956100 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f874947-e0c1-4e68-9c63-11016549a7db-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.466103 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.466678 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.697425 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1f874947-e0c1-4e68-9c63-11016549a7db","Type":"ContainerDied","Data":"1504002a2cfcfeaf90a24bb64fc62fc1c896d05f6713d4cb608bb46421cbe0c2"} Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.697467 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1504002a2cfcfeaf90a24bb64fc62fc1c896d05f6713d4cb608bb46421cbe0c2" Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.697473 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.766486 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.775489 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb825b-4e77-44f1-a14b-5a3660dd1799-metrics-certs\") pod \"network-metrics-daemon-d4t4l\" (UID: \"a0fb825b-4e77-44f1-a14b-5a3660dd1799\") " pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:52 crc kubenswrapper[4807]: I1205 12:08:52.855131 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4t4l" Dec 05 12:08:57 crc kubenswrapper[4807]: I1205 12:08:57.198744 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:08:57 crc kubenswrapper[4807]: I1205 12:08:57.311457 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-25wjt" Dec 05 12:08:58 crc kubenswrapper[4807]: I1205 12:08:58.010044 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:08:58 crc kubenswrapper[4807]: I1205 12:08:58.017304 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:09:08 crc kubenswrapper[4807]: I1205 12:09:08.249456 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ckgqz" Dec 05 12:09:13 crc kubenswrapper[4807]: E1205 12:09:13.244645 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 12:09:13 crc kubenswrapper[4807]: E1205 12:09:13.245148 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lm6jv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xsv4c_openshift-marketplace(2b219446-dca8-4aec-b300-92eb952d2a89): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:13 crc kubenswrapper[4807]: E1205 12:09:13.246504 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xsv4c" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.022557 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 12:09:18 crc kubenswrapper[4807]: E1205 12:09:18.022855 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3583f6c3-d1fd-4306-bda3-a0da3f00cb97" containerName="pruner" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.022872 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3583f6c3-d1fd-4306-bda3-a0da3f00cb97" containerName="pruner" Dec 05 12:09:18 crc kubenswrapper[4807]: E1205 12:09:18.022894 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f874947-e0c1-4e68-9c63-11016549a7db" containerName="pruner" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.022904 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f874947-e0c1-4e68-9c63-11016549a7db" containerName="pruner" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.023070 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3583f6c3-d1fd-4306-bda3-a0da3f00cb97" containerName="pruner" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.023087 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f874947-e0c1-4e68-9c63-11016549a7db" containerName="pruner" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.023638 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.028491 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.028664 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.039819 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.090502 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.090690 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.191632 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.192125 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.192207 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.225812 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.348274 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:18 crc kubenswrapper[4807]: I1205 12:09:18.691851 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 12:09:21 crc kubenswrapper[4807]: E1205 12:09:21.249070 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xsv4c" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" Dec 05 12:09:21 crc kubenswrapper[4807]: E1205 12:09:21.335182 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 12:09:21 crc kubenswrapper[4807]: E1205 12:09:21.335694 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j5z6f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9c8wq_openshift-marketplace(572bedd6-2920-481b-a356-c96fa8766456): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:21 crc kubenswrapper[4807]: E1205 12:09:21.336868 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9c8wq" podUID="572bedd6-2920-481b-a356-c96fa8766456" Dec 05 12:09:21 crc kubenswrapper[4807]: E1205 12:09:21.503862 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 12:09:21 crc kubenswrapper[4807]: E1205 12:09:21.504066 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7rcfp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-77zhj_openshift-marketplace(58c70b2e-9b4d-4820-b089-aec92a30ca58): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:21 crc kubenswrapper[4807]: E1205 12:09:21.505787 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-77zhj" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" Dec 05 12:09:22 crc kubenswrapper[4807]: I1205 12:09:22.466616 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:09:22 crc kubenswrapper[4807]: I1205 12:09:22.466684 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:09:22 crc kubenswrapper[4807]: E1205 12:09:22.869344 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-77zhj" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" Dec 05 12:09:22 crc kubenswrapper[4807]: E1205 12:09:22.869492 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9c8wq" podUID="572bedd6-2920-481b-a356-c96fa8766456" Dec 05 12:09:22 crc kubenswrapper[4807]: E1205 12:09:22.984784 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 12:09:22 crc kubenswrapper[4807]: E1205 12:09:22.985100 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljl7l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-sb9pm_openshift-marketplace(792cf8c2-f4b4-4956-955d-43541a2adcd3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:22 crc kubenswrapper[4807]: E1205 12:09:22.986249 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-sb9pm" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.012157 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.013077 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.025299 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.173470 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.173551 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23d20628-628f-414d-8841-20c56599d5e2-kube-api-access\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.173609 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-var-lock\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.274965 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-var-lock\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.275075 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.275106 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.275164 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23d20628-628f-414d-8841-20c56599d5e2-kube-api-access\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.275071 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-var-lock\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.295464 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23d20628-628f-414d-8841-20c56599d5e2-kube-api-access\") pod \"installer-9-crc\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:23 crc kubenswrapper[4807]: I1205 12:09:23.341591 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.454688 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-sb9pm" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.527676 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.527850 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7x7xv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ff68x_openshift-marketplace(02e14a82-3ab0-41f7-9ad4-5ca9ef73430d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.529246 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ff68x" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.573808 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.574215 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5t4gl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-927f9_openshift-marketplace(55077615-de5e-48c6-9be0-2f3c8bbabae7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.575810 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-927f9" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.586036 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.586194 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zfj9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4d4jf_openshift-marketplace(c9af52ad-e3d9-426f-9adf-a9880437d83f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.587385 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4d4jf" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.615043 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.615181 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gvks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jjbd7_openshift-marketplace(5f725110-9ece-41b3-a2af-2fa60bf2f649): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.616376 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jjbd7" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" Dec 05 12:09:24 crc kubenswrapper[4807]: I1205 12:09:24.705230 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 12:09:24 crc kubenswrapper[4807]: W1205 12:09:24.717399 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf3d364aa_bafb_49b8_8655_50e6a2ce3956.slice/crio-3f62b6d07c11587c6e8964626cbeb894c6777a3377ee3f00b63050550b7a77e0 WatchSource:0}: Error finding container 3f62b6d07c11587c6e8964626cbeb894c6777a3377ee3f00b63050550b7a77e0: Status 404 returned error can't find the container with id 3f62b6d07c11587c6e8964626cbeb894c6777a3377ee3f00b63050550b7a77e0 Dec 05 12:09:24 crc kubenswrapper[4807]: I1205 12:09:24.873146 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f3d364aa-bafb-49b8-8655-50e6a2ce3956","Type":"ContainerStarted","Data":"3f62b6d07c11587c6e8964626cbeb894c6777a3377ee3f00b63050550b7a77e0"} Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.874585 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jjbd7" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.875229 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-927f9" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.875248 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4d4jf" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" Dec 05 12:09:24 crc kubenswrapper[4807]: E1205 12:09:24.875284 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ff68x" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" Dec 05 12:09:24 crc kubenswrapper[4807]: I1205 12:09:24.960663 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-d4t4l"] Dec 05 12:09:24 crc kubenswrapper[4807]: I1205 12:09:24.960885 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 12:09:24 crc kubenswrapper[4807]: W1205 12:09:24.961951 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0fb825b_4e77_44f1_a14b_5a3660dd1799.slice/crio-93528f526dd50475fa4fa061273b117d2fd90997f8cc21ae19e5fda66a81654e WatchSource:0}: Error finding container 93528f526dd50475fa4fa061273b117d2fd90997f8cc21ae19e5fda66a81654e: Status 404 returned error can't find the container with id 93528f526dd50475fa4fa061273b117d2fd90997f8cc21ae19e5fda66a81654e Dec 05 12:09:24 crc kubenswrapper[4807]: W1205 12:09:24.965142 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod23d20628_628f_414d_8841_20c56599d5e2.slice/crio-b602194c133fab8d527d3ff5d32fa3346fe755d58bf7e5712ffefa327a273d91 WatchSource:0}: Error finding container b602194c133fab8d527d3ff5d32fa3346fe755d58bf7e5712ffefa327a273d91: Status 404 returned error can't find the container with id b602194c133fab8d527d3ff5d32fa3346fe755d58bf7e5712ffefa327a273d91 Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.897120 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" event={"ID":"a0fb825b-4e77-44f1-a14b-5a3660dd1799","Type":"ContainerStarted","Data":"efc0e5d9435ee9a6d95e47eb4e6130f6865fdba4681bc79d0409b7ec11504d40"} Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.897404 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" event={"ID":"a0fb825b-4e77-44f1-a14b-5a3660dd1799","Type":"ContainerStarted","Data":"215e2f0e7b9323fcd021a3642a32b0ca610c1ec3aa50f8a1a7a5093e16ff244e"} Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.897463 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-d4t4l" event={"ID":"a0fb825b-4e77-44f1-a14b-5a3660dd1799","Type":"ContainerStarted","Data":"93528f526dd50475fa4fa061273b117d2fd90997f8cc21ae19e5fda66a81654e"} Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.899563 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"23d20628-628f-414d-8841-20c56599d5e2","Type":"ContainerStarted","Data":"5259f5e62c0c811d4277c9a80c85f0b9eba63de1a832bcaf88c381397ae52c50"} Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.899588 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"23d20628-628f-414d-8841-20c56599d5e2","Type":"ContainerStarted","Data":"b602194c133fab8d527d3ff5d32fa3346fe755d58bf7e5712ffefa327a273d91"} Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.907088 4807 generic.go:334] "Generic (PLEG): container finished" podID="f3d364aa-bafb-49b8-8655-50e6a2ce3956" containerID="d322acc08d877edea2dde08d2ff1e67291fcf64a3f38da19d4317b5665c81527" exitCode=0 Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.907125 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f3d364aa-bafb-49b8-8655-50e6a2ce3956","Type":"ContainerDied","Data":"d322acc08d877edea2dde08d2ff1e67291fcf64a3f38da19d4317b5665c81527"} Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.915048 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-d4t4l" podStartSLOduration=174.91502617 podStartE2EDuration="2m54.91502617s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:09:25.910979185 +0000 UTC m=+195.404842454" watchObservedRunningTime="2025-12-05 12:09:25.91502617 +0000 UTC m=+195.408889449" Dec 05 12:09:25 crc kubenswrapper[4807]: I1205 12:09:25.931191 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.931172671 podStartE2EDuration="2.931172671s" podCreationTimestamp="2025-12-05 12:09:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:09:25.930035079 +0000 UTC m=+195.423898348" watchObservedRunningTime="2025-12-05 12:09:25.931172671 +0000 UTC m=+195.425035940" Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.191043 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.325730 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kube-api-access\") pod \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.325831 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kubelet-dir\") pod \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\" (UID: \"f3d364aa-bafb-49b8-8655-50e6a2ce3956\") " Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.325967 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f3d364aa-bafb-49b8-8655-50e6a2ce3956" (UID: "f3d364aa-bafb-49b8-8655-50e6a2ce3956"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.326134 4807 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.333347 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f3d364aa-bafb-49b8-8655-50e6a2ce3956" (UID: "f3d364aa-bafb-49b8-8655-50e6a2ce3956"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.427686 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3d364aa-bafb-49b8-8655-50e6a2ce3956-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.920276 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f3d364aa-bafb-49b8-8655-50e6a2ce3956","Type":"ContainerDied","Data":"3f62b6d07c11587c6e8964626cbeb894c6777a3377ee3f00b63050550b7a77e0"} Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.920307 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 12:09:27 crc kubenswrapper[4807]: I1205 12:09:27.920323 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f62b6d07c11587c6e8964626cbeb894c6777a3377ee3f00b63050550b7a77e0" Dec 05 12:09:35 crc kubenswrapper[4807]: I1205 12:09:35.954857 4807 generic.go:334] "Generic (PLEG): container finished" podID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerID="2d559b0cae43cf817b7973cd480bd5cb7c623c0f9f310177e177ce563af87c87" exitCode=0 Dec 05 12:09:35 crc kubenswrapper[4807]: I1205 12:09:35.954915 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77zhj" event={"ID":"58c70b2e-9b4d-4820-b089-aec92a30ca58","Type":"ContainerDied","Data":"2d559b0cae43cf817b7973cd480bd5cb7c623c0f9f310177e177ce563af87c87"} Dec 05 12:09:35 crc kubenswrapper[4807]: I1205 12:09:35.959821 4807 generic.go:334] "Generic (PLEG): container finished" podID="2b219446-dca8-4aec-b300-92eb952d2a89" containerID="4c204904a8921e521511678e7272e3a091e1052f464bfc06dd93da1afe10a9c3" exitCode=0 Dec 05 12:09:35 crc kubenswrapper[4807]: I1205 12:09:35.959856 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsv4c" event={"ID":"2b219446-dca8-4aec-b300-92eb952d2a89","Type":"ContainerDied","Data":"4c204904a8921e521511678e7272e3a091e1052f464bfc06dd93da1afe10a9c3"} Dec 05 12:09:36 crc kubenswrapper[4807]: I1205 12:09:36.967438 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77zhj" event={"ID":"58c70b2e-9b4d-4820-b089-aec92a30ca58","Type":"ContainerStarted","Data":"ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061"} Dec 05 12:09:36 crc kubenswrapper[4807]: I1205 12:09:36.969509 4807 generic.go:334] "Generic (PLEG): container finished" podID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerID="16ca63af1a3d6226cddced2d618ee976bad5b83bb3afdd6cd65e7e1e392019d4" exitCode=0 Dec 05 12:09:36 crc kubenswrapper[4807]: I1205 12:09:36.969597 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-927f9" event={"ID":"55077615-de5e-48c6-9be0-2f3c8bbabae7","Type":"ContainerDied","Data":"16ca63af1a3d6226cddced2d618ee976bad5b83bb3afdd6cd65e7e1e392019d4"} Dec 05 12:09:36 crc kubenswrapper[4807]: I1205 12:09:36.972042 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsv4c" event={"ID":"2b219446-dca8-4aec-b300-92eb952d2a89","Type":"ContainerStarted","Data":"a20b6c3dbdd61806689482174168b15fd1a791798889448104f7fca8f9b6af83"} Dec 05 12:09:37 crc kubenswrapper[4807]: I1205 12:09:37.006320 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-77zhj" podStartSLOduration=3.899571979 podStartE2EDuration="1m3.006304258s" podCreationTimestamp="2025-12-05 12:08:34 +0000 UTC" firstStartedPulling="2025-12-05 12:08:37.250698341 +0000 UTC m=+146.744561610" lastFinishedPulling="2025-12-05 12:09:36.35743062 +0000 UTC m=+205.851293889" observedRunningTime="2025-12-05 12:09:36.991100028 +0000 UTC m=+206.484963307" watchObservedRunningTime="2025-12-05 12:09:37.006304258 +0000 UTC m=+206.500167527" Dec 05 12:09:37 crc kubenswrapper[4807]: I1205 12:09:37.027984 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xsv4c" podStartSLOduration=2.690131706 podStartE2EDuration="1m3.027961573s" podCreationTimestamp="2025-12-05 12:08:34 +0000 UTC" firstStartedPulling="2025-12-05 12:08:36.042208317 +0000 UTC m=+145.536071586" lastFinishedPulling="2025-12-05 12:09:36.380038184 +0000 UTC m=+205.873901453" observedRunningTime="2025-12-05 12:09:37.023225417 +0000 UTC m=+206.517088706" watchObservedRunningTime="2025-12-05 12:09:37.027961573 +0000 UTC m=+206.521824842" Dec 05 12:09:37 crc kubenswrapper[4807]: I1205 12:09:37.983304 4807 generic.go:334] "Generic (PLEG): container finished" podID="572bedd6-2920-481b-a356-c96fa8766456" containerID="53552f960a25955ee166a8fc5f5e35dbfe1fcdf5bb616262a758ff44b97a5476" exitCode=0 Dec 05 12:09:37 crc kubenswrapper[4807]: I1205 12:09:37.983352 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9c8wq" event={"ID":"572bedd6-2920-481b-a356-c96fa8766456","Type":"ContainerDied","Data":"53552f960a25955ee166a8fc5f5e35dbfe1fcdf5bb616262a758ff44b97a5476"} Dec 05 12:09:37 crc kubenswrapper[4807]: I1205 12:09:37.985609 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-927f9" event={"ID":"55077615-de5e-48c6-9be0-2f3c8bbabae7","Type":"ContainerStarted","Data":"30fa918558f678fccbbd25be63737e328608cb478732caffe141ed136dad9c0e"} Dec 05 12:09:38 crc kubenswrapper[4807]: I1205 12:09:38.017644 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-927f9" podStartSLOduration=3.596543092 podStartE2EDuration="1m6.017622902s" podCreationTimestamp="2025-12-05 12:08:32 +0000 UTC" firstStartedPulling="2025-12-05 12:08:34.912715695 +0000 UTC m=+144.406578964" lastFinishedPulling="2025-12-05 12:09:37.333795495 +0000 UTC m=+206.827658774" observedRunningTime="2025-12-05 12:09:38.01618294 +0000 UTC m=+207.510046239" watchObservedRunningTime="2025-12-05 12:09:38.017622902 +0000 UTC m=+207.511486171" Dec 05 12:09:38 crc kubenswrapper[4807]: I1205 12:09:38.993140 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9c8wq" event={"ID":"572bedd6-2920-481b-a356-c96fa8766456","Type":"ContainerStarted","Data":"ec6d205c018c061bff0ca6c28357e23891d7f1a2f77b1eeb73790355d74cd903"} Dec 05 12:09:39 crc kubenswrapper[4807]: I1205 12:09:39.011931 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9c8wq" podStartSLOduration=4.030523421 podStartE2EDuration="1m4.011915154s" podCreationTimestamp="2025-12-05 12:08:35 +0000 UTC" firstStartedPulling="2025-12-05 12:08:38.380388908 +0000 UTC m=+147.874252177" lastFinishedPulling="2025-12-05 12:09:38.361780611 +0000 UTC m=+207.855643910" observedRunningTime="2025-12-05 12:09:39.009768242 +0000 UTC m=+208.503631511" watchObservedRunningTime="2025-12-05 12:09:39.011915154 +0000 UTC m=+208.505778423" Dec 05 12:09:40 crc kubenswrapper[4807]: I1205 12:09:40.000902 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d4jf" event={"ID":"c9af52ad-e3d9-426f-9adf-a9880437d83f","Type":"ContainerStarted","Data":"1b67e88b11c6a43a8a0a2bfc78c6f1fa45d8b9257497605bd462fd8d52619040"} Dec 05 12:09:40 crc kubenswrapper[4807]: I1205 12:09:40.002954 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff68x" event={"ID":"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d","Type":"ContainerStarted","Data":"93a3d68e6692e928df9718317121e20f9ce71ff56f89fadbcc307b31b2f88a66"} Dec 05 12:09:40 crc kubenswrapper[4807]: I1205 12:09:40.004766 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jjbd7" event={"ID":"5f725110-9ece-41b3-a2af-2fa60bf2f649","Type":"ContainerStarted","Data":"96cd8c604f0e7a77a13fffa11e1cd9106d43af3564b31416bb961b8d3bd45e73"} Dec 05 12:09:41 crc kubenswrapper[4807]: I1205 12:09:41.011676 4807 generic.go:334] "Generic (PLEG): container finished" podID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerID="1b67e88b11c6a43a8a0a2bfc78c6f1fa45d8b9257497605bd462fd8d52619040" exitCode=0 Dec 05 12:09:41 crc kubenswrapper[4807]: I1205 12:09:41.011801 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d4jf" event={"ID":"c9af52ad-e3d9-426f-9adf-a9880437d83f","Type":"ContainerDied","Data":"1b67e88b11c6a43a8a0a2bfc78c6f1fa45d8b9257497605bd462fd8d52619040"} Dec 05 12:09:41 crc kubenswrapper[4807]: I1205 12:09:41.015422 4807 generic.go:334] "Generic (PLEG): container finished" podID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerID="93a3d68e6692e928df9718317121e20f9ce71ff56f89fadbcc307b31b2f88a66" exitCode=0 Dec 05 12:09:41 crc kubenswrapper[4807]: I1205 12:09:41.015515 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff68x" event={"ID":"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d","Type":"ContainerDied","Data":"93a3d68e6692e928df9718317121e20f9ce71ff56f89fadbcc307b31b2f88a66"} Dec 05 12:09:41 crc kubenswrapper[4807]: I1205 12:09:41.017184 4807 generic.go:334] "Generic (PLEG): container finished" podID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerID="96cd8c604f0e7a77a13fffa11e1cd9106d43af3564b31416bb961b8d3bd45e73" exitCode=0 Dec 05 12:09:41 crc kubenswrapper[4807]: I1205 12:09:41.017241 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jjbd7" event={"ID":"5f725110-9ece-41b3-a2af-2fa60bf2f649","Type":"ContainerDied","Data":"96cd8c604f0e7a77a13fffa11e1cd9106d43af3564b31416bb961b8d3bd45e73"} Dec 05 12:09:43 crc kubenswrapper[4807]: I1205 12:09:43.043897 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:09:43 crc kubenswrapper[4807]: I1205 12:09:43.044815 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:09:43 crc kubenswrapper[4807]: I1205 12:09:43.120660 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.035342 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d4jf" event={"ID":"c9af52ad-e3d9-426f-9adf-a9880437d83f","Type":"ContainerStarted","Data":"9f915fda83de1bf39c93f25b297d80aa565698237cb2292c96519aabb44fbf9e"} Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.036884 4807 generic.go:334] "Generic (PLEG): container finished" podID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerID="b5dc690b5a3e2b43c9d44d533bb1fc92ea4660f26b2059505c5b4daf163cd668" exitCode=0 Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.036947 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb9pm" event={"ID":"792cf8c2-f4b4-4956-955d-43541a2adcd3","Type":"ContainerDied","Data":"b5dc690b5a3e2b43c9d44d533bb1fc92ea4660f26b2059505c5b4daf163cd668"} Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.038972 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff68x" event={"ID":"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d","Type":"ContainerStarted","Data":"a628b8ac905abee92f6633448587061d02761fb73f91e6b80279c11eb5711b60"} Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.043633 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jjbd7" event={"ID":"5f725110-9ece-41b3-a2af-2fa60bf2f649","Type":"ContainerStarted","Data":"885a31bae3ee941ac94a648c5246129d7d84dd3e4f606dab409e9c833a5a9881"} Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.058775 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4d4jf" podStartSLOduration=3.822321654 podStartE2EDuration="1m12.058757554s" podCreationTimestamp="2025-12-05 12:08:32 +0000 UTC" firstStartedPulling="2025-12-05 12:08:35.000896799 +0000 UTC m=+144.494760068" lastFinishedPulling="2025-12-05 12:09:43.237332699 +0000 UTC m=+212.731195968" observedRunningTime="2025-12-05 12:09:44.055043916 +0000 UTC m=+213.548907195" watchObservedRunningTime="2025-12-05 12:09:44.058757554 +0000 UTC m=+213.552620823" Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.103053 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.124329 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jjbd7" podStartSLOduration=4.422879824 podStartE2EDuration="1m9.124310119s" podCreationTimestamp="2025-12-05 12:08:35 +0000 UTC" firstStartedPulling="2025-12-05 12:08:38.380357227 +0000 UTC m=+147.874220496" lastFinishedPulling="2025-12-05 12:09:43.081787522 +0000 UTC m=+212.575650791" observedRunningTime="2025-12-05 12:09:44.106746221 +0000 UTC m=+213.600609490" watchObservedRunningTime="2025-12-05 12:09:44.124310119 +0000 UTC m=+213.618173398" Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.126639 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ff68x" podStartSLOduration=3.9733118320000003 podStartE2EDuration="1m12.126631556s" podCreationTimestamp="2025-12-05 12:08:32 +0000 UTC" firstStartedPulling="2025-12-05 12:08:34.939476357 +0000 UTC m=+144.433339636" lastFinishedPulling="2025-12-05 12:09:43.092796051 +0000 UTC m=+212.586659360" observedRunningTime="2025-12-05 12:09:44.123950918 +0000 UTC m=+213.617814187" watchObservedRunningTime="2025-12-05 12:09:44.126631556 +0000 UTC m=+213.620494825" Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.764822 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.764909 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:09:44 crc kubenswrapper[4807]: I1205 12:09:44.806899 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.049927 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb9pm" event={"ID":"792cf8c2-f4b4-4956-955d-43541a2adcd3","Type":"ContainerStarted","Data":"3359c8f7e08db096c425feec83e4fae1534bc7bc6f045e98107a19e73edaf57a"} Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.074840 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sb9pm" podStartSLOduration=4.731563235 podStartE2EDuration="1m13.074818355s" podCreationTimestamp="2025-12-05 12:08:32 +0000 UTC" firstStartedPulling="2025-12-05 12:08:36.114177308 +0000 UTC m=+145.608040577" lastFinishedPulling="2025-12-05 12:09:44.457432428 +0000 UTC m=+213.951295697" observedRunningTime="2025-12-05 12:09:45.072618332 +0000 UTC m=+214.566481611" watchObservedRunningTime="2025-12-05 12:09:45.074818355 +0000 UTC m=+214.568681624" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.096105 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.248705 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.248750 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.287001 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.860303 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.860371 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:09:45 crc kubenswrapper[4807]: I1205 12:09:45.901445 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:09:46 crc kubenswrapper[4807]: I1205 12:09:46.092205 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:09:46 crc kubenswrapper[4807]: I1205 12:09:46.094030 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:09:46 crc kubenswrapper[4807]: I1205 12:09:46.318741 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:09:46 crc kubenswrapper[4807]: I1205 12:09:46.318793 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:09:47 crc kubenswrapper[4807]: I1205 12:09:47.356012 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jjbd7" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="registry-server" probeResult="failure" output=< Dec 05 12:09:47 crc kubenswrapper[4807]: timeout: failed to connect service ":50051" within 1s Dec 05 12:09:47 crc kubenswrapper[4807]: > Dec 05 12:09:48 crc kubenswrapper[4807]: I1205 12:09:48.539298 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-77zhj"] Dec 05 12:09:48 crc kubenswrapper[4807]: I1205 12:09:48.539831 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-77zhj" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="registry-server" containerID="cri-o://ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061" gracePeriod=2 Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.465896 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.466000 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.466076 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.467118 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.467340 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a" gracePeriod=600 Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.741172 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.741265 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:09:52 crc kubenswrapper[4807]: I1205 12:09:52.790066 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:09:53 crc kubenswrapper[4807]: I1205 12:09:53.067855 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:09:53 crc kubenswrapper[4807]: I1205 12:09:53.067924 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:09:53 crc kubenswrapper[4807]: I1205 12:09:53.112101 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:09:53 crc kubenswrapper[4807]: I1205 12:09:53.257024 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:09:53 crc kubenswrapper[4807]: I1205 12:09:53.257114 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:09:53 crc kubenswrapper[4807]: I1205 12:09:53.293171 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:09:55 crc kubenswrapper[4807]: E1205 12:09:55.249298 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061 is running failed: container process not found" containerID="ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 12:09:55 crc kubenswrapper[4807]: E1205 12:09:55.250877 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061 is running failed: container process not found" containerID="ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 12:09:55 crc kubenswrapper[4807]: E1205 12:09:55.251227 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061 is running failed: container process not found" containerID="ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 12:09:55 crc kubenswrapper[4807]: E1205 12:09:55.251310 4807 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-77zhj" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="registry-server" Dec 05 12:09:56 crc kubenswrapper[4807]: I1205 12:09:56.356281 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:09:56 crc kubenswrapper[4807]: I1205 12:09:56.398766 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:09:56 crc kubenswrapper[4807]: I1205 12:09:56.823001 4807 generic.go:334] "Generic (PLEG): container finished" podID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerID="ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061" exitCode=0 Dec 05 12:09:56 crc kubenswrapper[4807]: I1205 12:09:56.824143 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77zhj" event={"ID":"58c70b2e-9b4d-4820-b089-aec92a30ca58","Type":"ContainerDied","Data":"ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061"} Dec 05 12:09:56 crc kubenswrapper[4807]: I1205 12:09:56.867139 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:09:56 crc kubenswrapper[4807]: I1205 12:09:56.877187 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.198134 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x4t9b"] Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.539067 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ff68x"] Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.810397 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.969995 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-catalog-content\") pod \"58c70b2e-9b4d-4820-b089-aec92a30ca58\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.970422 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-utilities\") pod \"58c70b2e-9b4d-4820-b089-aec92a30ca58\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.970469 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rcfp\" (UniqueName: \"kubernetes.io/projected/58c70b2e-9b4d-4820-b089-aec92a30ca58-kube-api-access-7rcfp\") pod \"58c70b2e-9b4d-4820-b089-aec92a30ca58\" (UID: \"58c70b2e-9b4d-4820-b089-aec92a30ca58\") " Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.972237 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-utilities" (OuterVolumeSpecName: "utilities") pod "58c70b2e-9b4d-4820-b089-aec92a30ca58" (UID: "58c70b2e-9b4d-4820-b089-aec92a30ca58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:09:57 crc kubenswrapper[4807]: I1205 12:09:57.986831 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58c70b2e-9b4d-4820-b089-aec92a30ca58" (UID: "58c70b2e-9b4d-4820-b089-aec92a30ca58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.071759 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.071793 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58c70b2e-9b4d-4820-b089-aec92a30ca58-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.117365 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c70b2e-9b4d-4820-b089-aec92a30ca58-kube-api-access-7rcfp" (OuterVolumeSpecName: "kube-api-access-7rcfp") pod "58c70b2e-9b4d-4820-b089-aec92a30ca58" (UID: "58c70b2e-9b4d-4820-b089-aec92a30ca58"). InnerVolumeSpecName "kube-api-access-7rcfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.173104 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rcfp\" (UniqueName: \"kubernetes.io/projected/58c70b2e-9b4d-4820-b089-aec92a30ca58-kube-api-access-7rcfp\") on node \"crc\" DevicePath \"\"" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.948805 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a" exitCode=0 Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.948896 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a"} Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.952342 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77zhj" event={"ID":"58c70b2e-9b4d-4820-b089-aec92a30ca58","Type":"ContainerDied","Data":"4636e9a1afbcd535191f9294fa6e0557f9cbb14ddb32d6a3a6415cf3cbb866ce"} Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.952459 4807 scope.go:117] "RemoveContainer" containerID="ba99c43996fb717b5bdff664a8f5fdba5e008b6d21c339259a406ac2f84c6061" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.952459 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77zhj" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.981748 4807 scope.go:117] "RemoveContainer" containerID="2d559b0cae43cf817b7973cd480bd5cb7c623c0f9f310177e177ce563af87c87" Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.989360 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-77zhj"] Dec 05 12:09:58 crc kubenswrapper[4807]: I1205 12:09:58.993606 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-77zhj"] Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.003679 4807 scope.go:117] "RemoveContainer" containerID="887faa743a261a6ce3c251075d938da2e81b467387a3e7228051b7a076c1fbc6" Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.014919 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.241557 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" path="/var/lib/kubelet/pods/58c70b2e-9b4d-4820-b089-aec92a30ca58/volumes" Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.738036 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sb9pm"] Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.944380 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jjbd7"] Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.958664 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"61fbdac8b8e46032c2f8db25d5c108f1020da47f0c78713ac4dad89d5fdc1819"} Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.959820 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ff68x" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="registry-server" containerID="cri-o://a628b8ac905abee92f6633448587061d02761fb73f91e6b80279c11eb5711b60" gracePeriod=2 Dec 05 12:09:59 crc kubenswrapper[4807]: I1205 12:09:59.960040 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jjbd7" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="registry-server" containerID="cri-o://885a31bae3ee941ac94a648c5246129d7d84dd3e4f606dab409e9c833a5a9881" gracePeriod=2 Dec 05 12:10:00 crc kubenswrapper[4807]: I1205 12:10:00.964450 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sb9pm" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="registry-server" containerID="cri-o://3359c8f7e08db096c425feec83e4fae1534bc7bc6f045e98107a19e73edaf57a" gracePeriod=2 Dec 05 12:10:01 crc kubenswrapper[4807]: I1205 12:10:01.972434 4807 generic.go:334] "Generic (PLEG): container finished" podID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerID="885a31bae3ee941ac94a648c5246129d7d84dd3e4f606dab409e9c833a5a9881" exitCode=0 Dec 05 12:10:01 crc kubenswrapper[4807]: I1205 12:10:01.972500 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jjbd7" event={"ID":"5f725110-9ece-41b3-a2af-2fa60bf2f649","Type":"ContainerDied","Data":"885a31bae3ee941ac94a648c5246129d7d84dd3e4f606dab409e9c833a5a9881"} Dec 05 12:10:01 crc kubenswrapper[4807]: I1205 12:10:01.976760 4807 generic.go:334] "Generic (PLEG): container finished" podID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerID="3359c8f7e08db096c425feec83e4fae1534bc7bc6f045e98107a19e73edaf57a" exitCode=0 Dec 05 12:10:01 crc kubenswrapper[4807]: I1205 12:10:01.976840 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb9pm" event={"ID":"792cf8c2-f4b4-4956-955d-43541a2adcd3","Type":"ContainerDied","Data":"3359c8f7e08db096c425feec83e4fae1534bc7bc6f045e98107a19e73edaf57a"} Dec 05 12:10:01 crc kubenswrapper[4807]: I1205 12:10:01.978863 4807 generic.go:334] "Generic (PLEG): container finished" podID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerID="a628b8ac905abee92f6633448587061d02761fb73f91e6b80279c11eb5711b60" exitCode=0 Dec 05 12:10:01 crc kubenswrapper[4807]: I1205 12:10:01.978934 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff68x" event={"ID":"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d","Type":"ContainerDied","Data":"a628b8ac905abee92f6633448587061d02761fb73f91e6b80279c11eb5711b60"} Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.036994 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.042191 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.125792 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gvks\" (UniqueName: \"kubernetes.io/projected/5f725110-9ece-41b3-a2af-2fa60bf2f649-kube-api-access-9gvks\") pod \"5f725110-9ece-41b3-a2af-2fa60bf2f649\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.127208 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x7xv\" (UniqueName: \"kubernetes.io/projected/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-kube-api-access-7x7xv\") pod \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.127291 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-catalog-content\") pod \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.127324 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-catalog-content\") pod \"5f725110-9ece-41b3-a2af-2fa60bf2f649\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.127347 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-utilities\") pod \"5f725110-9ece-41b3-a2af-2fa60bf2f649\" (UID: \"5f725110-9ece-41b3-a2af-2fa60bf2f649\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.127399 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-utilities\") pod \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\" (UID: \"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.128588 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-utilities" (OuterVolumeSpecName: "utilities") pod "02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" (UID: "02e14a82-3ab0-41f7-9ad4-5ca9ef73430d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.128949 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-utilities" (OuterVolumeSpecName: "utilities") pod "5f725110-9ece-41b3-a2af-2fa60bf2f649" (UID: "5f725110-9ece-41b3-a2af-2fa60bf2f649"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.134678 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-kube-api-access-7x7xv" (OuterVolumeSpecName: "kube-api-access-7x7xv") pod "02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" (UID: "02e14a82-3ab0-41f7-9ad4-5ca9ef73430d"). InnerVolumeSpecName "kube-api-access-7x7xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.134686 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f725110-9ece-41b3-a2af-2fa60bf2f649-kube-api-access-9gvks" (OuterVolumeSpecName: "kube-api-access-9gvks") pod "5f725110-9ece-41b3-a2af-2fa60bf2f649" (UID: "5f725110-9ece-41b3-a2af-2fa60bf2f649"). InnerVolumeSpecName "kube-api-access-9gvks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.178133 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.189204 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" (UID: "02e14a82-3ab0-41f7-9ad4-5ca9ef73430d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.228790 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.228844 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.228856 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.228869 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gvks\" (UniqueName: \"kubernetes.io/projected/5f725110-9ece-41b3-a2af-2fa60bf2f649-kube-api-access-9gvks\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.228883 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x7xv\" (UniqueName: \"kubernetes.io/projected/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d-kube-api-access-7x7xv\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.251766 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f725110-9ece-41b3-a2af-2fa60bf2f649" (UID: "5f725110-9ece-41b3-a2af-2fa60bf2f649"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.329786 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-catalog-content\") pod \"792cf8c2-f4b4-4956-955d-43541a2adcd3\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.330200 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-utilities\") pod \"792cf8c2-f4b4-4956-955d-43541a2adcd3\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.330331 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljl7l\" (UniqueName: \"kubernetes.io/projected/792cf8c2-f4b4-4956-955d-43541a2adcd3-kube-api-access-ljl7l\") pod \"792cf8c2-f4b4-4956-955d-43541a2adcd3\" (UID: \"792cf8c2-f4b4-4956-955d-43541a2adcd3\") " Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.331045 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f725110-9ece-41b3-a2af-2fa60bf2f649-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.332760 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-utilities" (OuterVolumeSpecName: "utilities") pod "792cf8c2-f4b4-4956-955d-43541a2adcd3" (UID: "792cf8c2-f4b4-4956-955d-43541a2adcd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.335625 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/792cf8c2-f4b4-4956-955d-43541a2adcd3-kube-api-access-ljl7l" (OuterVolumeSpecName: "kube-api-access-ljl7l") pod "792cf8c2-f4b4-4956-955d-43541a2adcd3" (UID: "792cf8c2-f4b4-4956-955d-43541a2adcd3"). InnerVolumeSpecName "kube-api-access-ljl7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.373519 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "792cf8c2-f4b4-4956-955d-43541a2adcd3" (UID: "792cf8c2-f4b4-4956-955d-43541a2adcd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.432498 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljl7l\" (UniqueName: \"kubernetes.io/projected/792cf8c2-f4b4-4956-955d-43541a2adcd3-kube-api-access-ljl7l\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.432603 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.432613 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792cf8c2-f4b4-4956-955d-43541a2adcd3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808604 4807 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808856 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808872 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808883 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808891 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808904 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808912 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808921 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808929 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808943 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808951 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808959 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808968 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808980 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d364aa-bafb-49b8-8655-50e6a2ce3956" containerName="pruner" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.808987 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d364aa-bafb-49b8-8655-50e6a2ce3956" containerName="pruner" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.808996 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809004 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.809012 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809020 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.809035 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809042 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.809053 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809061 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="extract-utilities" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.809070 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809076 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="extract-content" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.809085 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809092 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809227 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809245 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d364aa-bafb-49b8-8655-50e6a2ce3956" containerName="pruner" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809262 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809272 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809283 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c70b2e-9b4d-4820-b089-aec92a30ca58" containerName="registry-server" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809646 4807 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809901 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3" gracePeriod=15 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809957 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809995 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f" gracePeriod=15 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.809993 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928" gracePeriod=15 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.810013 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520" gracePeriod=15 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.810035 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5" gracePeriod=15 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812256 4807 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.812585 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812607 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.812626 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812639 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.812659 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812670 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.812689 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812701 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.812719 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812731 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.812745 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812755 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 12:10:02 crc kubenswrapper[4807]: E1205 12:10:02.812777 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812788 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812955 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.812992 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.813012 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.813028 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.813045 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.813372 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936643 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936694 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936717 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936738 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936755 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936773 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936807 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.936827 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.986400 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.988999 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.989977 4807 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928" exitCode=0 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.990031 4807 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f" exitCode=0 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.990040 4807 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5" exitCode=0 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.990051 4807 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520" exitCode=2 Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.990036 4807 scope.go:117] "RemoveContainer" containerID="054cddc67ccfde0918eac1f4265c30f06de1d8d3a4270fbf94af6911a0078aa0" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.993221 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb9pm" event={"ID":"792cf8c2-f4b4-4956-955d-43541a2adcd3","Type":"ContainerDied","Data":"795e5823cf490ed1f3e1137177461da5d8dbaaba90087d6d51f86e275eb97101"} Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.993283 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb9pm" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.994295 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.994558 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.995684 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff68x" event={"ID":"02e14a82-3ab0-41f7-9ad4-5ca9ef73430d","Type":"ContainerDied","Data":"67d1663936ed8e7a6751788f559cffe8016c3046e58ce693ee0c434891d06c36"} Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.995790 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff68x" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.997048 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.997410 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:02 crc kubenswrapper[4807]: I1205 12:10:02.999700 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.001659 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jjbd7" event={"ID":"5f725110-9ece-41b3-a2af-2fa60bf2f649","Type":"ContainerDied","Data":"62f0ddbc29a2fcc498edc8bebb8f8c61d3a20fefc1936f9d86cddfefde7e736e"} Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.001722 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jjbd7" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.002437 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.002777 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.003280 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.006721 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037672 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037802 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037842 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037856 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037909 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037868 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037979 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.038006 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037985 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.037978 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.038044 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.038075 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.038111 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.038153 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.038156 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.038126 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.081805 4807 scope.go:117] "RemoveContainer" containerID="3359c8f7e08db096c425feec83e4fae1534bc7bc6f045e98107a19e73edaf57a" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.088381 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.088829 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.089131 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.089580 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.090148 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.090397 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.090721 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.090998 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.098104 4807 scope.go:117] "RemoveContainer" containerID="b5dc690b5a3e2b43c9d44d533bb1fc92ea4660f26b2059505c5b4daf163cd668" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.136094 4807 scope.go:117] "RemoveContainer" containerID="da7366947a208d15882969e83f229edeb5b4386c2e52bb09e87424910646bdb9" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.149228 4807 scope.go:117] "RemoveContainer" containerID="a628b8ac905abee92f6633448587061d02761fb73f91e6b80279c11eb5711b60" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.176417 4807 scope.go:117] "RemoveContainer" containerID="93a3d68e6692e928df9718317121e20f9ce71ff56f89fadbcc307b31b2f88a66" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.188470 4807 scope.go:117] "RemoveContainer" containerID="182f2ef6a5d2e4bc76f7fa70c6c0c3363e831452e7f92540dc5e054341813fa6" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.207553 4807 scope.go:117] "RemoveContainer" containerID="885a31bae3ee941ac94a648c5246129d7d84dd3e4f606dab409e9c833a5a9881" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.220240 4807 scope.go:117] "RemoveContainer" containerID="96cd8c604f0e7a77a13fffa11e1cd9106d43af3564b31416bb961b8d3bd45e73" Dec 05 12:10:03 crc kubenswrapper[4807]: I1205 12:10:03.237327 4807 scope.go:117] "RemoveContainer" containerID="df330a45426c14b0f6fe534e5471637d7fc15f910f47ad581c7464126eb5a658" Dec 05 12:10:03 crc kubenswrapper[4807]: E1205 12:10:03.263025 4807 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" volumeName="registry-storage" Dec 05 12:10:04 crc kubenswrapper[4807]: I1205 12:10:04.009660 4807 generic.go:334] "Generic (PLEG): container finished" podID="23d20628-628f-414d-8841-20c56599d5e2" containerID="5259f5e62c0c811d4277c9a80c85f0b9eba63de1a832bcaf88c381397ae52c50" exitCode=0 Dec 05 12:10:04 crc kubenswrapper[4807]: I1205 12:10:04.009770 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"23d20628-628f-414d-8841-20c56599d5e2","Type":"ContainerDied","Data":"5259f5e62c0c811d4277c9a80c85f0b9eba63de1a832bcaf88c381397ae52c50"} Dec 05 12:10:04 crc kubenswrapper[4807]: I1205 12:10:04.010751 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:04 crc kubenswrapper[4807]: I1205 12:10:04.011386 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:04 crc kubenswrapper[4807]: I1205 12:10:04.011600 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:04 crc kubenswrapper[4807]: I1205 12:10:04.011910 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:04 crc kubenswrapper[4807]: I1205 12:10:04.018038 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.308279 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.310395 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.311234 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.312091 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.312680 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.470729 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-var-lock\") pod \"23d20628-628f-414d-8841-20c56599d5e2\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.471289 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-kubelet-dir\") pod \"23d20628-628f-414d-8841-20c56599d5e2\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.470882 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-var-lock" (OuterVolumeSpecName: "var-lock") pod "23d20628-628f-414d-8841-20c56599d5e2" (UID: "23d20628-628f-414d-8841-20c56599d5e2"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.471453 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "23d20628-628f-414d-8841-20c56599d5e2" (UID: "23d20628-628f-414d-8841-20c56599d5e2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.471473 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23d20628-628f-414d-8841-20c56599d5e2-kube-api-access\") pod \"23d20628-628f-414d-8841-20c56599d5e2\" (UID: \"23d20628-628f-414d-8841-20c56599d5e2\") " Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.471793 4807 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.471813 4807 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23d20628-628f-414d-8841-20c56599d5e2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.480117 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23d20628-628f-414d-8841-20c56599d5e2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "23d20628-628f-414d-8841-20c56599d5e2" (UID: "23d20628-628f-414d-8841-20c56599d5e2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.572871 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23d20628-628f-414d-8841-20c56599d5e2-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:05 crc kubenswrapper[4807]: E1205 12:10:05.595741 4807 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: E1205 12:10:05.596376 4807 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: E1205 12:10:05.596851 4807 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: E1205 12:10:05.597342 4807 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: E1205 12:10:05.597658 4807 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.597697 4807 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 12:10:05 crc kubenswrapper[4807]: E1205 12:10:05.597986 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="200ms" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.658047 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.659182 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.659701 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.659974 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.660394 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.660634 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.660883 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.774823 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.774886 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.775030 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.775063 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.775083 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.775166 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.775379 4807 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.775411 4807 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:05 crc kubenswrapper[4807]: I1205 12:10:05.775426 4807 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:05 crc kubenswrapper[4807]: E1205 12:10:05.798685 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="400ms" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.029626 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"23d20628-628f-414d-8841-20c56599d5e2","Type":"ContainerDied","Data":"b602194c133fab8d527d3ff5d32fa3346fe755d58bf7e5712ffefa327a273d91"} Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.029664 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b602194c133fab8d527d3ff5d32fa3346fe755d58bf7e5712ffefa327a273d91" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.029653 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.031642 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.032129 4807 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3" exitCode=0 Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.032176 4807 scope.go:117] "RemoveContainer" containerID="4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.032283 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.044039 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.044358 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.044675 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.045092 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.045402 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.047100 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.047367 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.047605 4807 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.047798 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.047990 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.050502 4807 scope.go:117] "RemoveContainer" containerID="5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.063815 4807 scope.go:117] "RemoveContainer" containerID="bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.076442 4807 scope.go:117] "RemoveContainer" containerID="1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.087479 4807 scope.go:117] "RemoveContainer" containerID="d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.101763 4807 scope.go:117] "RemoveContainer" containerID="0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.123032 4807 scope.go:117] "RemoveContainer" containerID="4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.123608 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\": container with ID starting with 4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928 not found: ID does not exist" containerID="4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.123660 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928"} err="failed to get container status \"4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\": rpc error: code = NotFound desc = could not find container \"4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928\": container with ID starting with 4d21ebca7c603d54fce0b7db0c4eb0f1603035a8369691c2ba993f0b90e1d928 not found: ID does not exist" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.123686 4807 scope.go:117] "RemoveContainer" containerID="5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.124085 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\": container with ID starting with 5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f not found: ID does not exist" containerID="5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.124138 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f"} err="failed to get container status \"5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\": rpc error: code = NotFound desc = could not find container \"5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f\": container with ID starting with 5d88d8603753492b0d6be602d86847cea07daf8b7b33db6b0549c27c9e284f2f not found: ID does not exist" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.124490 4807 scope.go:117] "RemoveContainer" containerID="bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.124812 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\": container with ID starting with bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5 not found: ID does not exist" containerID="bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.124830 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5"} err="failed to get container status \"bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\": rpc error: code = NotFound desc = could not find container \"bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5\": container with ID starting with bc780039558b0a1dacd09c781a4016c61489ca6807d8bb6bd3c02081522fd1f5 not found: ID does not exist" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.124845 4807 scope.go:117] "RemoveContainer" containerID="1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.125264 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\": container with ID starting with 1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520 not found: ID does not exist" containerID="1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.125284 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520"} err="failed to get container status \"1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\": rpc error: code = NotFound desc = could not find container \"1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520\": container with ID starting with 1bb9d9cbb4e54e144cbdbbac2e6d32354a965c6e5e44bc037f6a587ab0359520 not found: ID does not exist" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.125297 4807 scope.go:117] "RemoveContainer" containerID="d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.125497 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\": container with ID starting with d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3 not found: ID does not exist" containerID="d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.125518 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3"} err="failed to get container status \"d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\": rpc error: code = NotFound desc = could not find container \"d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3\": container with ID starting with d21f6e1e2eca141ecbe7013879038c94ed9975c0ce82cc7af8db2cf487cdfaa3 not found: ID does not exist" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.125552 4807 scope.go:117] "RemoveContainer" containerID="0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.125757 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\": container with ID starting with 0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa not found: ID does not exist" containerID="0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa" Dec 05 12:10:06 crc kubenswrapper[4807]: I1205 12:10:06.125782 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa"} err="failed to get container status \"0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\": rpc error: code = NotFound desc = could not find container \"0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa\": container with ID starting with 0af426b7673ff7aae564254586591f82a8fe8ef525eb83f07d3ffc53de8967fa not found: ID does not exist" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.199805 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="800ms" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.925438 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:10:06Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:10:06Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:10:06Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T12:10:06Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.926090 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.926596 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.926959 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.927208 4807 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:06 crc kubenswrapper[4807]: E1205 12:10:06.927225 4807 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 12:10:07 crc kubenswrapper[4807]: E1205 12:10:07.001590 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="1.6s" Dec 05 12:10:07 crc kubenswrapper[4807]: I1205 12:10:07.246509 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 05 12:10:07 crc kubenswrapper[4807]: E1205 12:10:07.844457 4807 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:07 crc kubenswrapper[4807]: I1205 12:10:07.845387 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:07 crc kubenswrapper[4807]: W1205 12:10:07.883022 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-5e30eb402cb5c15aa22704d0402bc2180915364944ed8e4a3e157ecf8cda132c WatchSource:0}: Error finding container 5e30eb402cb5c15aa22704d0402bc2180915364944ed8e4a3e157ecf8cda132c: Status 404 returned error can't find the container with id 5e30eb402cb5c15aa22704d0402bc2180915364944ed8e4a3e157ecf8cda132c Dec 05 12:10:07 crc kubenswrapper[4807]: E1205 12:10:07.887098 4807 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e507e6c29a59f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 12:10:07.886583199 +0000 UTC m=+237.380446468,LastTimestamp:2025-12-05 12:10:07.886583199 +0000 UTC m=+237.380446468,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 12:10:08 crc kubenswrapper[4807]: I1205 12:10:08.046180 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"5e30eb402cb5c15aa22704d0402bc2180915364944ed8e4a3e157ecf8cda132c"} Dec 05 12:10:08 crc kubenswrapper[4807]: E1205 12:10:08.602048 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="3.2s" Dec 05 12:10:09 crc kubenswrapper[4807]: I1205 12:10:09.056066 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d"} Dec 05 12:10:09 crc kubenswrapper[4807]: E1205 12:10:09.057273 4807 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:09 crc kubenswrapper[4807]: I1205 12:10:09.057327 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:09 crc kubenswrapper[4807]: I1205 12:10:09.057720 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:09 crc kubenswrapper[4807]: I1205 12:10:09.058006 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:09 crc kubenswrapper[4807]: I1205 12:10:09.058266 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:10 crc kubenswrapper[4807]: E1205 12:10:10.063369 4807 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:11 crc kubenswrapper[4807]: I1205 12:10:11.238763 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:11 crc kubenswrapper[4807]: I1205 12:10:11.239114 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:11 crc kubenswrapper[4807]: I1205 12:10:11.239351 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:11 crc kubenswrapper[4807]: I1205 12:10:11.239959 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:11 crc kubenswrapper[4807]: E1205 12:10:11.803485 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="6.4s" Dec 05 12:10:14 crc kubenswrapper[4807]: E1205 12:10:14.539096 4807 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e507e6c29a59f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 12:10:07.886583199 +0000 UTC m=+237.380446468,LastTimestamp:2025-12-05 12:10:07.886583199 +0000 UTC m=+237.380446468,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.234749 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.235719 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.236055 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.236576 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.237021 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.249093 4807 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.249119 4807 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:17 crc kubenswrapper[4807]: E1205 12:10:17.249512 4807 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:17 crc kubenswrapper[4807]: I1205 12:10:17.250036 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.118804 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.119115 4807 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35" exitCode=1 Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.119192 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35"} Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.119698 4807 scope.go:117] "RemoveContainer" containerID="fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.120277 4807 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.120690 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121079 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121227 4807 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f426bc914a70f04d2ebb3ca9c4e404114629fadf68a4a98836dca707c5f1acff" exitCode=0 Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121267 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f426bc914a70f04d2ebb3ca9c4e404114629fadf68a4a98836dca707c5f1acff"} Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121333 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7de2c420acfd362886b5c97da9090e82e3783cdfd3e4c53b2821d8f405e4ab79"} Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121560 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121742 4807 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121771 4807 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.121808 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: E1205 12:10:18.122195 4807 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.122248 4807 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.122555 4807 status_manager.go:851] "Failed to get status for pod" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" pod="openshift-marketplace/certified-operators-sb9pm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-sb9pm\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.122786 4807 status_manager.go:851] "Failed to get status for pod" podUID="23d20628-628f-414d-8841-20c56599d5e2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.123077 4807 status_manager.go:851] "Failed to get status for pod" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" pod="openshift-marketplace/redhat-operators-jjbd7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-jjbd7\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: I1205 12:10:18.123337 4807 status_manager.go:851] "Failed to get status for pod" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" pod="openshift-marketplace/community-operators-ff68x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ff68x\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 05 12:10:18 crc kubenswrapper[4807]: E1205 12:10:18.204197 4807 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="7s" Dec 05 12:10:19 crc kubenswrapper[4807]: I1205 12:10:19.129805 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 12:10:19 crc kubenswrapper[4807]: I1205 12:10:19.130188 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"24b88096658c348918045022129d8ed73e20e53d9f0784b2d1cd61cad022d185"} Dec 05 12:10:19 crc kubenswrapper[4807]: I1205 12:10:19.134156 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e9b1a7f52f27495bde034318f3be90f292395b6bdf2031be9e8fafc90b44f14c"} Dec 05 12:10:19 crc kubenswrapper[4807]: I1205 12:10:19.134212 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"45f22c144e93d1def02fac0425d011a7bb3de30f7c0d1cb7106b767b6df8278e"} Dec 05 12:10:19 crc kubenswrapper[4807]: I1205 12:10:19.134223 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"467ee6054be2cd7088d84fa32acd65aca6b27246bccd7358f6583608f3be343a"} Dec 05 12:10:19 crc kubenswrapper[4807]: I1205 12:10:19.134235 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5b6a1307f83fbacb23be683e72d4850f34967f902e70feb1259dfe373802b58a"} Dec 05 12:10:20 crc kubenswrapper[4807]: I1205 12:10:20.142966 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dbb16dcf43263ffe1b022940aea2593630230777769e7118fa581191c2d627d4"} Dec 05 12:10:20 crc kubenswrapper[4807]: I1205 12:10:20.143198 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:20 crc kubenswrapper[4807]: I1205 12:10:20.143338 4807 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:20 crc kubenswrapper[4807]: I1205 12:10:20.143364 4807 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.010371 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.010538 4807 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.010749 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.112473 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.223151 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" podUID="c63ee192-3315-4176-9654-8497d5ba9fd7" containerName="oauth-openshift" containerID="cri-o://7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef" gracePeriod=15 Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.251155 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.251344 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.257052 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.538771 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594487 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-login\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594570 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-error\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594603 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsbbd\" (UniqueName: \"kubernetes.io/projected/c63ee192-3315-4176-9654-8497d5ba9fd7-kube-api-access-nsbbd\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594671 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-ocp-branding-template\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594725 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-serving-cert\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594752 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-dir\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594794 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-idp-0-file-data\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594822 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-service-ca\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594838 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-policies\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594894 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-router-certs\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594906 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594921 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-session\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.594972 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-trusted-ca-bundle\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.595002 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-provider-selection\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.595038 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-cliconfig\") pod \"c63ee192-3315-4176-9654-8497d5ba9fd7\" (UID: \"c63ee192-3315-4176-9654-8497d5ba9fd7\") " Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.595272 4807 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.595645 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.595730 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.596110 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.596666 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.602448 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.602744 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c63ee192-3315-4176-9654-8497d5ba9fd7-kube-api-access-nsbbd" (OuterVolumeSpecName: "kube-api-access-nsbbd") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "kube-api-access-nsbbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.602769 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.603170 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.603748 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.604047 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.604386 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.604552 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.609955 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c63ee192-3315-4176-9654-8497d5ba9fd7" (UID: "c63ee192-3315-4176-9654-8497d5ba9fd7"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696039 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696086 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696099 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsbbd\" (UniqueName: \"kubernetes.io/projected/c63ee192-3315-4176-9654-8497d5ba9fd7-kube-api-access-nsbbd\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696113 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696127 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696139 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696150 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696162 4807 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696174 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696185 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696198 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696210 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:22 crc kubenswrapper[4807]: I1205 12:10:22.696222 4807 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c63ee192-3315-4176-9654-8497d5ba9fd7-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:23 crc kubenswrapper[4807]: I1205 12:10:23.157983 4807 generic.go:334] "Generic (PLEG): container finished" podID="c63ee192-3315-4176-9654-8497d5ba9fd7" containerID="7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef" exitCode=0 Dec 05 12:10:23 crc kubenswrapper[4807]: I1205 12:10:23.158025 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" event={"ID":"c63ee192-3315-4176-9654-8497d5ba9fd7","Type":"ContainerDied","Data":"7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef"} Dec 05 12:10:23 crc kubenswrapper[4807]: I1205 12:10:23.158056 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" event={"ID":"c63ee192-3315-4176-9654-8497d5ba9fd7","Type":"ContainerDied","Data":"ca41fbc46910ffcaee583675c6bb391edae416ecdf5e0c39957194a32b2a477b"} Dec 05 12:10:23 crc kubenswrapper[4807]: I1205 12:10:23.158085 4807 scope.go:117] "RemoveContainer" containerID="7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef" Dec 05 12:10:23 crc kubenswrapper[4807]: I1205 12:10:23.158089 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x4t9b" Dec 05 12:10:23 crc kubenswrapper[4807]: I1205 12:10:23.175411 4807 scope.go:117] "RemoveContainer" containerID="7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef" Dec 05 12:10:23 crc kubenswrapper[4807]: E1205 12:10:23.175949 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef\": container with ID starting with 7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef not found: ID does not exist" containerID="7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef" Dec 05 12:10:23 crc kubenswrapper[4807]: I1205 12:10:23.176098 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef"} err="failed to get container status \"7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef\": rpc error: code = NotFound desc = could not find container \"7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef\": container with ID starting with 7b3bc1f2787256d4622ec030029865921d7a09a735b825c7ad168c2bf09324ef not found: ID does not exist" Dec 05 12:10:24 crc kubenswrapper[4807]: E1205 12:10:24.816342 4807 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 05 12:10:25 crc kubenswrapper[4807]: E1205 12:10:25.044887 4807 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 05 12:10:25 crc kubenswrapper[4807]: I1205 12:10:25.172404 4807 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:25 crc kubenswrapper[4807]: E1205 12:10:25.521270 4807 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 05 12:10:25 crc kubenswrapper[4807]: E1205 12:10:25.882327 4807 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 05 12:10:26 crc kubenswrapper[4807]: I1205 12:10:26.174790 4807 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:26 crc kubenswrapper[4807]: I1205 12:10:26.174830 4807 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:26 crc kubenswrapper[4807]: I1205 12:10:26.179704 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:26 crc kubenswrapper[4807]: I1205 12:10:26.182762 4807 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0b7afcb5-4364-4412-8c93-e499b9bf9ed3" Dec 05 12:10:27 crc kubenswrapper[4807]: I1205 12:10:27.179668 4807 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:27 crc kubenswrapper[4807]: I1205 12:10:27.179710 4807 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a27a2e4-e6d4-4de5-b18a-a7e806069a49" Dec 05 12:10:31 crc kubenswrapper[4807]: I1205 12:10:31.245459 4807 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0b7afcb5-4364-4412-8c93-e499b9bf9ed3" Dec 05 12:10:32 crc kubenswrapper[4807]: I1205 12:10:32.010440 4807 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 12:10:32 crc kubenswrapper[4807]: I1205 12:10:32.010496 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 12:10:33 crc kubenswrapper[4807]: I1205 12:10:33.945701 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.001680 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.222920 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.382157 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.475934 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.507047 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.648954 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.686820 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 12:10:35 crc kubenswrapper[4807]: I1205 12:10:35.758852 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 12:10:36 crc kubenswrapper[4807]: I1205 12:10:36.117672 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 12:10:36 crc kubenswrapper[4807]: I1205 12:10:36.166182 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 12:10:36 crc kubenswrapper[4807]: I1205 12:10:36.448054 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 12:10:36 crc kubenswrapper[4807]: I1205 12:10:36.897020 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 12:10:36 crc kubenswrapper[4807]: I1205 12:10:36.904899 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 12:10:36 crc kubenswrapper[4807]: I1205 12:10:36.930445 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.042934 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.045371 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.097139 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.098839 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.146902 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.166302 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.308432 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.354881 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.438440 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.469147 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.756160 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.868885 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.918413 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.968350 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.979787 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 12:10:37 crc kubenswrapper[4807]: I1205 12:10:37.994920 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.008366 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.177269 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.178260 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.186292 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.217368 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.263553 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.285709 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.308415 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.335801 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.338921 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.446421 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.473842 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.476560 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.513252 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.585169 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.727183 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.743330 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.754420 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.779002 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.839784 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.839946 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.841372 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 12:10:38 crc kubenswrapper[4807]: I1205 12:10:38.843208 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.011423 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.024557 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.042924 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.060718 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.100791 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.139275 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.189449 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.200462 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.231675 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.250058 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.286544 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.302398 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.362312 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.433033 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.451177 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.567404 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.617824 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.744185 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.759016 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.780080 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.790135 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 12:10:39 crc kubenswrapper[4807]: I1205 12:10:39.916620 4807 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.007588 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.146315 4807 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.197287 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.251048 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.260008 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.302893 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.413676 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.442043 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.547860 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.721867 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.771243 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.836426 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.909053 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.918950 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 12:10:40 crc kubenswrapper[4807]: I1205 12:10:40.997557 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.202351 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.394324 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.469346 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.504420 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.529200 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.627012 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.682411 4807 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.702328 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 12:10:41 crc kubenswrapper[4807]: I1205 12:10:41.884020 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.010888 4807 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.010969 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.011034 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.011726 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"24b88096658c348918045022129d8ed73e20e53d9f0784b2d1cd61cad022d185"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.011857 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://24b88096658c348918045022129d8ed73e20e53d9f0784b2d1cd61cad022d185" gracePeriod=30 Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.035074 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.036909 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.056066 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.174699 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.187554 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.235133 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.240381 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.274927 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.332391 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.338207 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.342431 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.385993 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.439564 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.567089 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.617294 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.644000 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.647189 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.690081 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.750582 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.772342 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.774436 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.912705 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.938926 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 12:10:42 crc kubenswrapper[4807]: I1205 12:10:42.996927 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.012387 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.057664 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.195511 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.204495 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.226259 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.245961 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.255554 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.274112 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.278599 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.281001 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.501192 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.524301 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.603097 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.614180 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.710360 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.743905 4807 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.782871 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.828333 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.880835 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.914968 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.933644 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 12:10:43 crc kubenswrapper[4807]: I1205 12:10:43.939002 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.075809 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.172517 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.191433 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.193059 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.259502 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.454206 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.595204 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.730507 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.748353 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.848139 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.961337 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 12:10:44 crc kubenswrapper[4807]: I1205 12:10:44.974329 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.019792 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.137012 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.143304 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.201195 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.266232 4807 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.271624 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sb9pm","openshift-marketplace/redhat-operators-jjbd7","openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-x4t9b","openshift-marketplace/community-operators-ff68x"] Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.271733 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.276820 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.283166 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.290505 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.290480093 podStartE2EDuration="20.290480093s" podCreationTimestamp="2025-12-05 12:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:10:45.286129509 +0000 UTC m=+274.779992788" watchObservedRunningTime="2025-12-05 12:10:45.290480093 +0000 UTC m=+274.784343372" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.356482 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.363233 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.406308 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.445982 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.515059 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.563629 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.715675 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.727227 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.729790 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.734186 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.744426 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.746030 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.805494 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.887070 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.895127 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 12:10:45 crc kubenswrapper[4807]: I1205 12:10:45.949434 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.003286 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.090689 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.117458 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.207203 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.295838 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.308984 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.314549 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.338506 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.342963 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.383448 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.395417 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.415150 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.493126 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.604312 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.709355 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.743666 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.751199 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.774086 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.791346 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.957138 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 12:10:46 crc kubenswrapper[4807]: I1205 12:10:46.987143 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.004639 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.097918 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.247450 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02e14a82-3ab0-41f7-9ad4-5ca9ef73430d" path="/var/lib/kubelet/pods/02e14a82-3ab0-41f7-9ad4-5ca9ef73430d/volumes" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.249164 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f725110-9ece-41b3-a2af-2fa60bf2f649" path="/var/lib/kubelet/pods/5f725110-9ece-41b3-a2af-2fa60bf2f649/volumes" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.250508 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="792cf8c2-f4b4-4956-955d-43541a2adcd3" path="/var/lib/kubelet/pods/792cf8c2-f4b4-4956-955d-43541a2adcd3/volumes" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.253149 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c63ee192-3315-4176-9654-8497d5ba9fd7" path="/var/lib/kubelet/pods/c63ee192-3315-4176-9654-8497d5ba9fd7/volumes" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.284246 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.284261 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.310820 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.377679 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.426287 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.440758 4807 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.440972 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d" gracePeriod=5 Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.488907 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.566616 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.604777 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.645862 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.658469 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.761981 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.813086 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 12:10:47 crc kubenswrapper[4807]: I1205 12:10:47.964561 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 12:10:48 crc kubenswrapper[4807]: I1205 12:10:48.002422 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 12:10:48 crc kubenswrapper[4807]: I1205 12:10:48.155299 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 12:10:48 crc kubenswrapper[4807]: I1205 12:10:48.196921 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 12:10:48 crc kubenswrapper[4807]: I1205 12:10:48.277593 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 12:10:48 crc kubenswrapper[4807]: I1205 12:10:48.427101 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 12:10:48 crc kubenswrapper[4807]: I1205 12:10:48.624997 4807 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 12:10:48 crc kubenswrapper[4807]: I1205 12:10:48.883912 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 12:10:49 crc kubenswrapper[4807]: I1205 12:10:49.109034 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 12:10:49 crc kubenswrapper[4807]: I1205 12:10:49.228652 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 12:10:49 crc kubenswrapper[4807]: I1205 12:10:49.511623 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 12:10:49 crc kubenswrapper[4807]: I1205 12:10:49.650335 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 12:10:49 crc kubenswrapper[4807]: I1205 12:10:49.667982 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 12:10:49 crc kubenswrapper[4807]: I1205 12:10:49.693988 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 12:10:49 crc kubenswrapper[4807]: I1205 12:10:49.809321 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 12:10:50 crc kubenswrapper[4807]: I1205 12:10:50.131210 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 12:10:50 crc kubenswrapper[4807]: I1205 12:10:50.257762 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 12:10:50 crc kubenswrapper[4807]: I1205 12:10:50.289615 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 12:10:50 crc kubenswrapper[4807]: I1205 12:10:50.608610 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 12:10:50 crc kubenswrapper[4807]: I1205 12:10:50.952191 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 12:10:51 crc kubenswrapper[4807]: I1205 12:10:51.608644 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.008880 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.009256 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171263 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171354 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171389 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171464 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171508 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171535 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171658 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171702 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.171810 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.172198 4807 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.172221 4807 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.172231 4807 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.172259 4807 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.182388 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.241650 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.273701 4807 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.311455 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.311618 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.311676 4807 scope.go:117] "RemoveContainer" containerID="9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.311515 4807 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d" exitCode=137 Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.326003 4807 scope.go:117] "RemoveContainer" containerID="9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d" Dec 05 12:10:53 crc kubenswrapper[4807]: E1205 12:10:53.326385 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d\": container with ID starting with 9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d not found: ID does not exist" containerID="9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d" Dec 05 12:10:53 crc kubenswrapper[4807]: I1205 12:10:53.326440 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d"} err="failed to get container status \"9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d\": rpc error: code = NotFound desc = could not find container \"9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d\": container with ID starting with 9c4fde6ade7d85425dfafa113d40d3e72ee4125e660a5b80b283a93db4d1ad2d not found: ID does not exist" Dec 05 12:11:04 crc kubenswrapper[4807]: I1205 12:11:04.372287 4807 generic.go:334] "Generic (PLEG): container finished" podID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerID="f4ab674188cc0f8873a0882d7a969dc85235e8c5d9409617ec7dfdda9c4a1b16" exitCode=0 Dec 05 12:11:04 crc kubenswrapper[4807]: I1205 12:11:04.372373 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" event={"ID":"71ce4423-09bb-43f1-a7a9-86143ea5dd9d","Type":"ContainerDied","Data":"f4ab674188cc0f8873a0882d7a969dc85235e8c5d9409617ec7dfdda9c4a1b16"} Dec 05 12:11:04 crc kubenswrapper[4807]: I1205 12:11:04.373392 4807 scope.go:117] "RemoveContainer" containerID="f4ab674188cc0f8873a0882d7a969dc85235e8c5d9409617ec7dfdda9c4a1b16" Dec 05 12:11:05 crc kubenswrapper[4807]: I1205 12:11:05.381999 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" event={"ID":"71ce4423-09bb-43f1-a7a9-86143ea5dd9d","Type":"ContainerStarted","Data":"0fbf59fc0fd565cbd9ea441a4740f73c7e54bc81488acc206b0aad55003e8517"} Dec 05 12:11:05 crc kubenswrapper[4807]: I1205 12:11:05.382750 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:11:05 crc kubenswrapper[4807]: I1205 12:11:05.389634 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:11:12 crc kubenswrapper[4807]: I1205 12:11:12.421226 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 05 12:11:12 crc kubenswrapper[4807]: I1205 12:11:12.424402 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 12:11:12 crc kubenswrapper[4807]: I1205 12:11:12.424450 4807 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="24b88096658c348918045022129d8ed73e20e53d9f0784b2d1cd61cad022d185" exitCode=137 Dec 05 12:11:12 crc kubenswrapper[4807]: I1205 12:11:12.424483 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"24b88096658c348918045022129d8ed73e20e53d9f0784b2d1cd61cad022d185"} Dec 05 12:11:12 crc kubenswrapper[4807]: I1205 12:11:12.424517 4807 scope.go:117] "RemoveContainer" containerID="fd4558ba3e98a3709f1edbd457f9b546dacf613c80fb71a7a218086a27aaab35" Dec 05 12:11:13 crc kubenswrapper[4807]: I1205 12:11:13.432185 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 05 12:11:13 crc kubenswrapper[4807]: I1205 12:11:13.433768 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3ff0db0e80f57510af12a00803160688f0bd64de126e42b5f6be4108e518bfdf"} Dec 05 12:11:22 crc kubenswrapper[4807]: I1205 12:11:22.010595 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:11:22 crc kubenswrapper[4807]: I1205 12:11:22.015909 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:11:22 crc kubenswrapper[4807]: I1205 12:11:22.112987 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:11:22 crc kubenswrapper[4807]: I1205 12:11:22.119491 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 12:11:24 crc kubenswrapper[4807]: I1205 12:11:24.817445 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.064332 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57864598c6-x28kw"] Dec 05 12:11:27 crc kubenswrapper[4807]: E1205 12:11:27.064902 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.064917 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 12:11:27 crc kubenswrapper[4807]: E1205 12:11:27.064931 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d20628-628f-414d-8841-20c56599d5e2" containerName="installer" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.064939 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d20628-628f-414d-8841-20c56599d5e2" containerName="installer" Dec 05 12:11:27 crc kubenswrapper[4807]: E1205 12:11:27.064956 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c63ee192-3315-4176-9654-8497d5ba9fd7" containerName="oauth-openshift" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.064964 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c63ee192-3315-4176-9654-8497d5ba9fd7" containerName="oauth-openshift" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.065074 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c63ee192-3315-4176-9654-8497d5ba9fd7" containerName="oauth-openshift" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.065091 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.065103 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d20628-628f-414d-8841-20c56599d5e2" containerName="installer" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.065514 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070079 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070298 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070402 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070445 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070470 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070456 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070673 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070836 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070847 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.070970 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.071220 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.071329 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.080035 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.081952 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.086877 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57864598c6-x28kw"] Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.091645 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191352 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-session\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191416 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191451 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/201daa99-df28-4f30-827b-a21ac7cdf1dd-audit-dir\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191563 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191607 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slm7t\" (UniqueName: \"kubernetes.io/projected/201daa99-df28-4f30-827b-a21ac7cdf1dd-kube-api-access-slm7t\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191727 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-audit-policies\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191751 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-login\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191774 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-router-certs\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191802 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191845 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-error\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191867 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191891 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.191966 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-service-ca\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.192019 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293424 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293560 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-error\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293592 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293619 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293645 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-service-ca\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293689 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293724 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293751 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-session\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293796 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/201daa99-df28-4f30-827b-a21ac7cdf1dd-audit-dir\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293835 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293870 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slm7t\" (UniqueName: \"kubernetes.io/projected/201daa99-df28-4f30-827b-a21ac7cdf1dd-kube-api-access-slm7t\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293937 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-audit-policies\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293964 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-router-certs\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.293987 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-login\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.294507 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/201daa99-df28-4f30-827b-a21ac7cdf1dd-audit-dir\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.295133 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-audit-policies\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.295609 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.295705 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-service-ca\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.295889 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.300173 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.300246 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.300296 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-router-certs\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.301722 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-session\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.306852 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-error\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.307104 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-login\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.307295 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.308826 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/201daa99-df28-4f30-827b-a21ac7cdf1dd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.310994 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slm7t\" (UniqueName: \"kubernetes.io/projected/201daa99-df28-4f30-827b-a21ac7cdf1dd-kube-api-access-slm7t\") pod \"oauth-openshift-57864598c6-x28kw\" (UID: \"201daa99-df28-4f30-827b-a21ac7cdf1dd\") " pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.400996 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:27 crc kubenswrapper[4807]: I1205 12:11:27.578702 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57864598c6-x28kw"] Dec 05 12:11:28 crc kubenswrapper[4807]: I1205 12:11:28.514942 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" event={"ID":"201daa99-df28-4f30-827b-a21ac7cdf1dd","Type":"ContainerStarted","Data":"7f2e80074bc7621fb0753a9b408614c18003cfb6ea803372e5ea23a1a3ffc514"} Dec 05 12:11:28 crc kubenswrapper[4807]: I1205 12:11:28.515574 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:28 crc kubenswrapper[4807]: I1205 12:11:28.515592 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" event={"ID":"201daa99-df28-4f30-827b-a21ac7cdf1dd","Type":"ContainerStarted","Data":"ffd149ec159d205b24d5a4b1283d5065b460fc90400be99ad0956515b4ed2fbb"} Dec 05 12:11:28 crc kubenswrapper[4807]: I1205 12:11:28.543416 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" podStartSLOduration=91.543394849 podStartE2EDuration="1m31.543394849s" podCreationTimestamp="2025-12-05 12:09:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:11:28.53165956 +0000 UTC m=+318.025522839" watchObservedRunningTime="2025-12-05 12:11:28.543394849 +0000 UTC m=+318.037258118" Dec 05 12:11:28 crc kubenswrapper[4807]: I1205 12:11:28.974326 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57864598c6-x28kw" Dec 05 12:11:36 crc kubenswrapper[4807]: I1205 12:11:36.943918 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h2v8s"] Dec 05 12:11:36 crc kubenswrapper[4807]: I1205 12:11:36.944344 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" podUID="b6e5fb53-aec3-445a-827f-3673c9a18f2b" containerName="controller-manager" containerID="cri-o://0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b" gracePeriod=30 Dec 05 12:11:36 crc kubenswrapper[4807]: I1205 12:11:36.956024 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr"] Dec 05 12:11:36 crc kubenswrapper[4807]: I1205 12:11:36.956278 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" podUID="37a88100-9cdf-482d-8963-fd6215a3ca67" containerName="route-controller-manager" containerID="cri-o://f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f" gracePeriod=30 Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.408080 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.458544 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.516409 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkpgv\" (UniqueName: \"kubernetes.io/projected/37a88100-9cdf-482d-8963-fd6215a3ca67-kube-api-access-rkpgv\") pod \"37a88100-9cdf-482d-8963-fd6215a3ca67\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.516836 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37a88100-9cdf-482d-8963-fd6215a3ca67-serving-cert\") pod \"37a88100-9cdf-482d-8963-fd6215a3ca67\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.517852 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-proxy-ca-bundles\") pod \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.518417 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc55m\" (UniqueName: \"kubernetes.io/projected/b6e5fb53-aec3-445a-827f-3673c9a18f2b-kube-api-access-bc55m\") pod \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.518483 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-config\") pod \"37a88100-9cdf-482d-8963-fd6215a3ca67\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.518517 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-client-ca\") pod \"37a88100-9cdf-482d-8963-fd6215a3ca67\" (UID: \"37a88100-9cdf-482d-8963-fd6215a3ca67\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.518574 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-client-ca\") pod \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.518618 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-config\") pod \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.518645 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e5fb53-aec3-445a-827f-3673c9a18f2b-serving-cert\") pod \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\" (UID: \"b6e5fb53-aec3-445a-827f-3673c9a18f2b\") " Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.519065 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b6e5fb53-aec3-445a-827f-3673c9a18f2b" (UID: "b6e5fb53-aec3-445a-827f-3673c9a18f2b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.519465 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-client-ca" (OuterVolumeSpecName: "client-ca") pod "b6e5fb53-aec3-445a-827f-3673c9a18f2b" (UID: "b6e5fb53-aec3-445a-827f-3673c9a18f2b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.519712 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-client-ca" (OuterVolumeSpecName: "client-ca") pod "37a88100-9cdf-482d-8963-fd6215a3ca67" (UID: "37a88100-9cdf-482d-8963-fd6215a3ca67"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.519874 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-config" (OuterVolumeSpecName: "config") pod "37a88100-9cdf-482d-8963-fd6215a3ca67" (UID: "37a88100-9cdf-482d-8963-fd6215a3ca67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.520097 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-config" (OuterVolumeSpecName: "config") pod "b6e5fb53-aec3-445a-827f-3673c9a18f2b" (UID: "b6e5fb53-aec3-445a-827f-3673c9a18f2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.522802 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e5fb53-aec3-445a-827f-3673c9a18f2b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b6e5fb53-aec3-445a-827f-3673c9a18f2b" (UID: "b6e5fb53-aec3-445a-827f-3673c9a18f2b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.523033 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e5fb53-aec3-445a-827f-3673c9a18f2b-kube-api-access-bc55m" (OuterVolumeSpecName: "kube-api-access-bc55m") pod "b6e5fb53-aec3-445a-827f-3673c9a18f2b" (UID: "b6e5fb53-aec3-445a-827f-3673c9a18f2b"). InnerVolumeSpecName "kube-api-access-bc55m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.523304 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37a88100-9cdf-482d-8963-fd6215a3ca67-kube-api-access-rkpgv" (OuterVolumeSpecName: "kube-api-access-rkpgv") pod "37a88100-9cdf-482d-8963-fd6215a3ca67" (UID: "37a88100-9cdf-482d-8963-fd6215a3ca67"). InnerVolumeSpecName "kube-api-access-rkpgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.523997 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37a88100-9cdf-482d-8963-fd6215a3ca67-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "37a88100-9cdf-482d-8963-fd6215a3ca67" (UID: "37a88100-9cdf-482d-8963-fd6215a3ca67"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.573420 4807 generic.go:334] "Generic (PLEG): container finished" podID="37a88100-9cdf-482d-8963-fd6215a3ca67" containerID="f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f" exitCode=0 Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.573472 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" event={"ID":"37a88100-9cdf-482d-8963-fd6215a3ca67","Type":"ContainerDied","Data":"f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f"} Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.573502 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.573554 4807 scope.go:117] "RemoveContainer" containerID="f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.573539 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr" event={"ID":"37a88100-9cdf-482d-8963-fd6215a3ca67","Type":"ContainerDied","Data":"29af8102beae524743b0c72ddbca9abb229ce0b42c95356957e851f96419bb71"} Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.575049 4807 generic.go:334] "Generic (PLEG): container finished" podID="b6e5fb53-aec3-445a-827f-3673c9a18f2b" containerID="0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b" exitCode=0 Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.575095 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.575107 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" event={"ID":"b6e5fb53-aec3-445a-827f-3673c9a18f2b","Type":"ContainerDied","Data":"0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b"} Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.575130 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h2v8s" event={"ID":"b6e5fb53-aec3-445a-827f-3673c9a18f2b","Type":"ContainerDied","Data":"1f776d94f3da1771d062b22a54d114f0e9c185396a8d135204af5a607cf855eb"} Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.588722 4807 scope.go:117] "RemoveContainer" containerID="f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f" Dec 05 12:11:37 crc kubenswrapper[4807]: E1205 12:11:37.589588 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f\": container with ID starting with f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f not found: ID does not exist" containerID="f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.589625 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f"} err="failed to get container status \"f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f\": rpc error: code = NotFound desc = could not find container \"f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f\": container with ID starting with f5379596a8daf50b60debb6592d00001f2e67a96bb7db1a6e142bdd53c068c3f not found: ID does not exist" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.589672 4807 scope.go:117] "RemoveContainer" containerID="0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.601280 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr"] Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.606223 4807 scope.go:117] "RemoveContainer" containerID="0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b" Dec 05 12:11:37 crc kubenswrapper[4807]: E1205 12:11:37.606703 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b\": container with ID starting with 0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b not found: ID does not exist" containerID="0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.606745 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b"} err="failed to get container status \"0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b\": rpc error: code = NotFound desc = could not find container \"0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b\": container with ID starting with 0c06ad0426667213e8a047402dc6c50335c6830228e861df75daea2b9007220b not found: ID does not exist" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.609126 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gk5fr"] Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.615138 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h2v8s"] Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619616 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619645 4807 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37a88100-9cdf-482d-8963-fd6215a3ca67-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619657 4807 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619665 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e5fb53-aec3-445a-827f-3673c9a18f2b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619677 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619688 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkpgv\" (UniqueName: \"kubernetes.io/projected/37a88100-9cdf-482d-8963-fd6215a3ca67-kube-api-access-rkpgv\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619698 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37a88100-9cdf-482d-8963-fd6215a3ca67-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619706 4807 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6e5fb53-aec3-445a-827f-3673c9a18f2b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.619714 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc55m\" (UniqueName: \"kubernetes.io/projected/b6e5fb53-aec3-445a-827f-3673c9a18f2b-kube-api-access-bc55m\") on node \"crc\" DevicePath \"\"" Dec 05 12:11:37 crc kubenswrapper[4807]: I1205 12:11:37.620591 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h2v8s"] Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.233089 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94"] Dec 05 12:11:38 crc kubenswrapper[4807]: E1205 12:11:38.233386 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e5fb53-aec3-445a-827f-3673c9a18f2b" containerName="controller-manager" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.233400 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e5fb53-aec3-445a-827f-3673c9a18f2b" containerName="controller-manager" Dec 05 12:11:38 crc kubenswrapper[4807]: E1205 12:11:38.233412 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37a88100-9cdf-482d-8963-fd6215a3ca67" containerName="route-controller-manager" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.233421 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="37a88100-9cdf-482d-8963-fd6215a3ca67" containerName="route-controller-manager" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.233546 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e5fb53-aec3-445a-827f-3673c9a18f2b" containerName="controller-manager" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.233562 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="37a88100-9cdf-482d-8963-fd6215a3ca67" containerName="route-controller-manager" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.233990 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.235831 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.236160 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.236286 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.236465 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5898c9b65b-vwcl9"] Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.236595 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.236744 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.237086 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.237290 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.240603 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.240686 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.240789 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.243741 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.243926 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.248847 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94"] Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.249651 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.251557 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.255605 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5898c9b65b-vwcl9"] Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.327700 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-client-ca\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328012 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-config\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328173 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d896b724-8139-4a4a-8e04-833d994abba1-serving-cert\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328303 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbvrk\" (UniqueName: \"kubernetes.io/projected/a49e4323-289a-4433-9b78-33da0d2e3f33-kube-api-access-pbvrk\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328423 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d896b724-8139-4a4a-8e04-833d994abba1-client-ca\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328570 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-proxy-ca-bundles\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328707 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d896b724-8139-4a4a-8e04-833d994abba1-config\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328862 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nchl\" (UniqueName: \"kubernetes.io/projected/d896b724-8139-4a4a-8e04-833d994abba1-kube-api-access-5nchl\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.328979 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49e4323-289a-4433-9b78-33da0d2e3f33-serving-cert\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430325 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-client-ca\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430714 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-config\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430763 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d896b724-8139-4a4a-8e04-833d994abba1-serving-cert\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430788 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbvrk\" (UniqueName: \"kubernetes.io/projected/a49e4323-289a-4433-9b78-33da0d2e3f33-kube-api-access-pbvrk\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430813 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d896b724-8139-4a4a-8e04-833d994abba1-client-ca\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430852 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-proxy-ca-bundles\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430874 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d896b724-8139-4a4a-8e04-833d994abba1-config\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430893 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nchl\" (UniqueName: \"kubernetes.io/projected/d896b724-8139-4a4a-8e04-833d994abba1-kube-api-access-5nchl\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.430921 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49e4323-289a-4433-9b78-33da0d2e3f33-serving-cert\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.432369 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d896b724-8139-4a4a-8e04-833d994abba1-client-ca\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.432408 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d896b724-8139-4a4a-8e04-833d994abba1-config\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.432582 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-client-ca\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.433084 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-proxy-ca-bundles\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.433592 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-config\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.435322 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d896b724-8139-4a4a-8e04-833d994abba1-serving-cert\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.443108 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49e4323-289a-4433-9b78-33da0d2e3f33-serving-cert\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.448349 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbvrk\" (UniqueName: \"kubernetes.io/projected/a49e4323-289a-4433-9b78-33da0d2e3f33-kube-api-access-pbvrk\") pod \"controller-manager-5898c9b65b-vwcl9\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.453379 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nchl\" (UniqueName: \"kubernetes.io/projected/d896b724-8139-4a4a-8e04-833d994abba1-kube-api-access-5nchl\") pod \"route-controller-manager-845d8c8ddb-ndv94\" (UID: \"d896b724-8139-4a4a-8e04-833d994abba1\") " pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.561107 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.569738 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:38 crc kubenswrapper[4807]: I1205 12:11:38.984072 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5898c9b65b-vwcl9"] Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.026776 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94"] Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.241577 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37a88100-9cdf-482d-8963-fd6215a3ca67" path="/var/lib/kubelet/pods/37a88100-9cdf-482d-8963-fd6215a3ca67/volumes" Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.242239 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e5fb53-aec3-445a-827f-3673c9a18f2b" path="/var/lib/kubelet/pods/b6e5fb53-aec3-445a-827f-3673c9a18f2b/volumes" Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.595450 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" event={"ID":"a49e4323-289a-4433-9b78-33da0d2e3f33","Type":"ContainerStarted","Data":"0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76"} Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.595715 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" event={"ID":"a49e4323-289a-4433-9b78-33da0d2e3f33","Type":"ContainerStarted","Data":"dc3e848d93c65f66a691d4decd08ee8aa8a4695b36ba3be707f8609484c7603c"} Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.595735 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.596921 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" event={"ID":"d896b724-8139-4a4a-8e04-833d994abba1","Type":"ContainerStarted","Data":"6c6a81938ae0a77f1632da13c14dc4f95b77f3476d6b710f066000919a91e333"} Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.596951 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" event={"ID":"d896b724-8139-4a4a-8e04-833d994abba1","Type":"ContainerStarted","Data":"981bf6ac39f55c7f4b0f22b5e3fd1d2e682600d1414d25d55b884979ec9e9882"} Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.597756 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.602635 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.642182 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" podStartSLOduration=3.642159914 podStartE2EDuration="3.642159914s" podCreationTimestamp="2025-12-05 12:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:11:39.627215861 +0000 UTC m=+329.121079150" watchObservedRunningTime="2025-12-05 12:11:39.642159914 +0000 UTC m=+329.136023183" Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.646042 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" Dec 05 12:11:39 crc kubenswrapper[4807]: I1205 12:11:39.665006 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-845d8c8ddb-ndv94" podStartSLOduration=2.664984523 podStartE2EDuration="2.664984523s" podCreationTimestamp="2025-12-05 12:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:11:39.659683354 +0000 UTC m=+329.153546633" watchObservedRunningTime="2025-12-05 12:11:39.664984523 +0000 UTC m=+329.158847812" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.321833 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9t6mh"] Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.324075 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.338484 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9t6mh"] Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.521809 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-registry-certificates\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.521870 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.521903 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-bound-sa-token\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.522146 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-trusted-ca\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.522219 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbt7d\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-kube-api-access-rbt7d\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.522255 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.522290 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-registry-tls\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.522382 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.542905 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.623664 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-trusted-ca\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.623736 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbt7d\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-kube-api-access-rbt7d\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.623767 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.623790 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-registry-tls\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.623832 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.623861 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-registry-certificates\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.623901 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-bound-sa-token\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.624651 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.625454 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-registry-certificates\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.625608 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-trusted-ca\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.636751 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.636790 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-registry-tls\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.641447 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbt7d\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-kube-api-access-rbt7d\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.641488 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/46b205c2-a2b6-492e-ab0b-1a1499e89b5b-bound-sa-token\") pod \"image-registry-66df7c8f76-9t6mh\" (UID: \"46b205c2-a2b6-492e-ab0b-1a1499e89b5b\") " pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:02 crc kubenswrapper[4807]: I1205 12:12:02.940806 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:03 crc kubenswrapper[4807]: I1205 12:12:03.337595 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9t6mh"] Dec 05 12:12:03 crc kubenswrapper[4807]: W1205 12:12:03.340737 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46b205c2_a2b6_492e_ab0b_1a1499e89b5b.slice/crio-152cde0458fbc56e8e04f55ff1026445298a6ded501b99069b5ab2b2412f78c1 WatchSource:0}: Error finding container 152cde0458fbc56e8e04f55ff1026445298a6ded501b99069b5ab2b2412f78c1: Status 404 returned error can't find the container with id 152cde0458fbc56e8e04f55ff1026445298a6ded501b99069b5ab2b2412f78c1 Dec 05 12:12:03 crc kubenswrapper[4807]: I1205 12:12:03.707313 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" event={"ID":"46b205c2-a2b6-492e-ab0b-1a1499e89b5b","Type":"ContainerStarted","Data":"152cde0458fbc56e8e04f55ff1026445298a6ded501b99069b5ab2b2412f78c1"} Dec 05 12:12:04 crc kubenswrapper[4807]: I1205 12:12:04.715450 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" event={"ID":"46b205c2-a2b6-492e-ab0b-1a1499e89b5b","Type":"ContainerStarted","Data":"02ae6ddbe4974f9e3e645d4230b0f709c2226b1501abc748f12f35dbec326aa4"} Dec 05 12:12:04 crc kubenswrapper[4807]: I1205 12:12:04.716909 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:04 crc kubenswrapper[4807]: I1205 12:12:04.742773 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" podStartSLOduration=2.742753494 podStartE2EDuration="2.742753494s" podCreationTimestamp="2025-12-05 12:12:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:12:04.738992914 +0000 UTC m=+354.232856193" watchObservedRunningTime="2025-12-05 12:12:04.742753494 +0000 UTC m=+354.236616763" Dec 05 12:12:11 crc kubenswrapper[4807]: I1205 12:12:11.702577 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5898c9b65b-vwcl9"] Dec 05 12:12:11 crc kubenswrapper[4807]: I1205 12:12:11.703333 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" podUID="a49e4323-289a-4433-9b78-33da0d2e3f33" containerName="controller-manager" containerID="cri-o://0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76" gracePeriod=30 Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.147221 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.266396 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-config\") pod \"a49e4323-289a-4433-9b78-33da0d2e3f33\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.266437 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbvrk\" (UniqueName: \"kubernetes.io/projected/a49e4323-289a-4433-9b78-33da0d2e3f33-kube-api-access-pbvrk\") pod \"a49e4323-289a-4433-9b78-33da0d2e3f33\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.266513 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-client-ca\") pod \"a49e4323-289a-4433-9b78-33da0d2e3f33\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.266558 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49e4323-289a-4433-9b78-33da0d2e3f33-serving-cert\") pod \"a49e4323-289a-4433-9b78-33da0d2e3f33\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.266601 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-proxy-ca-bundles\") pod \"a49e4323-289a-4433-9b78-33da0d2e3f33\" (UID: \"a49e4323-289a-4433-9b78-33da0d2e3f33\") " Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.267406 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-config" (OuterVolumeSpecName: "config") pod "a49e4323-289a-4433-9b78-33da0d2e3f33" (UID: "a49e4323-289a-4433-9b78-33da0d2e3f33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.267429 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-client-ca" (OuterVolumeSpecName: "client-ca") pod "a49e4323-289a-4433-9b78-33da0d2e3f33" (UID: "a49e4323-289a-4433-9b78-33da0d2e3f33"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.267418 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a49e4323-289a-4433-9b78-33da0d2e3f33" (UID: "a49e4323-289a-4433-9b78-33da0d2e3f33"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.274666 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a49e4323-289a-4433-9b78-33da0d2e3f33-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a49e4323-289a-4433-9b78-33da0d2e3f33" (UID: "a49e4323-289a-4433-9b78-33da0d2e3f33"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.274739 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a49e4323-289a-4433-9b78-33da0d2e3f33-kube-api-access-pbvrk" (OuterVolumeSpecName: "kube-api-access-pbvrk") pod "a49e4323-289a-4433-9b78-33da0d2e3f33" (UID: "a49e4323-289a-4433-9b78-33da0d2e3f33"). InnerVolumeSpecName "kube-api-access-pbvrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.368128 4807 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.368164 4807 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49e4323-289a-4433-9b78-33da0d2e3f33-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.368175 4807 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.368189 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49e4323-289a-4433-9b78-33da0d2e3f33-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.368202 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbvrk\" (UniqueName: \"kubernetes.io/projected/a49e4323-289a-4433-9b78-33da0d2e3f33-kube-api-access-pbvrk\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.770381 4807 generic.go:334] "Generic (PLEG): container finished" podID="a49e4323-289a-4433-9b78-33da0d2e3f33" containerID="0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76" exitCode=0 Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.770424 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" event={"ID":"a49e4323-289a-4433-9b78-33da0d2e3f33","Type":"ContainerDied","Data":"0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76"} Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.770451 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" event={"ID":"a49e4323-289a-4433-9b78-33da0d2e3f33","Type":"ContainerDied","Data":"dc3e848d93c65f66a691d4decd08ee8aa8a4695b36ba3be707f8609484c7603c"} Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.770480 4807 scope.go:117] "RemoveContainer" containerID="0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.770616 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5898c9b65b-vwcl9" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.788773 4807 scope.go:117] "RemoveContainer" containerID="0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76" Dec 05 12:12:12 crc kubenswrapper[4807]: E1205 12:12:12.789207 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76\": container with ID starting with 0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76 not found: ID does not exist" containerID="0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.789243 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76"} err="failed to get container status \"0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76\": rpc error: code = NotFound desc = could not find container \"0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76\": container with ID starting with 0f370c5741d75af75c0ff4a5d488fbe3547724c30e41d9a1004bf99481f58d76 not found: ID does not exist" Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.800060 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5898c9b65b-vwcl9"] Dec 05 12:12:12 crc kubenswrapper[4807]: I1205 12:12:12.803336 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5898c9b65b-vwcl9"] Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.241438 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a49e4323-289a-4433-9b78-33da0d2e3f33" path="/var/lib/kubelet/pods/a49e4323-289a-4433-9b78-33da0d2e3f33/volumes" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.254878 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-87c8f44b5-jd6tr"] Dec 05 12:12:13 crc kubenswrapper[4807]: E1205 12:12:13.255100 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a49e4323-289a-4433-9b78-33da0d2e3f33" containerName="controller-manager" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.255113 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a49e4323-289a-4433-9b78-33da0d2e3f33" containerName="controller-manager" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.255196 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a49e4323-289a-4433-9b78-33da0d2e3f33" containerName="controller-manager" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.255605 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.257594 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.257888 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.258203 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.258412 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.258632 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.260212 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.265257 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.268164 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87c8f44b5-jd6tr"] Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.385699 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlfb8\" (UniqueName: \"kubernetes.io/projected/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-kube-api-access-xlfb8\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.386374 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-client-ca\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.386488 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-config\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.386546 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-proxy-ca-bundles\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.386940 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-serving-cert\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.488969 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-serving-cert\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.489105 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlfb8\" (UniqueName: \"kubernetes.io/projected/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-kube-api-access-xlfb8\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.489190 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-client-ca\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.489261 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-config\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.489323 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-proxy-ca-bundles\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.490351 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-client-ca\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.491136 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-config\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.491152 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-proxy-ca-bundles\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.493543 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-serving-cert\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.508686 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlfb8\" (UniqueName: \"kubernetes.io/projected/6c036adf-afd0-41dd-b7ae-ef26d99cbbec-kube-api-access-xlfb8\") pod \"controller-manager-87c8f44b5-jd6tr\" (UID: \"6c036adf-afd0-41dd-b7ae-ef26d99cbbec\") " pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.576179 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:13 crc kubenswrapper[4807]: I1205 12:12:13.993508 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87c8f44b5-jd6tr"] Dec 05 12:12:14 crc kubenswrapper[4807]: I1205 12:12:14.783790 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" event={"ID":"6c036adf-afd0-41dd-b7ae-ef26d99cbbec","Type":"ContainerStarted","Data":"13ba599542284127b2ba30ee787d60177d9add71e1992a06a95942e104653623"} Dec 05 12:12:14 crc kubenswrapper[4807]: I1205 12:12:14.784296 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:14 crc kubenswrapper[4807]: I1205 12:12:14.784312 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" event={"ID":"6c036adf-afd0-41dd-b7ae-ef26d99cbbec","Type":"ContainerStarted","Data":"17ef7d948abc3d7e81e70fff460235174034b7ffc52e84f342564c60328b0214"} Dec 05 12:12:14 crc kubenswrapper[4807]: I1205 12:12:14.788333 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" Dec 05 12:12:14 crc kubenswrapper[4807]: I1205 12:12:14.807512 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-87c8f44b5-jd6tr" podStartSLOduration=3.807494429 podStartE2EDuration="3.807494429s" podCreationTimestamp="2025-12-05 12:12:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:12:14.801747128 +0000 UTC m=+364.295610407" watchObservedRunningTime="2025-12-05 12:12:14.807494429 +0000 UTC m=+364.301357698" Dec 05 12:12:22 crc kubenswrapper[4807]: I1205 12:12:22.466415 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:12:22 crc kubenswrapper[4807]: I1205 12:12:22.467026 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:12:22 crc kubenswrapper[4807]: I1205 12:12:22.945825 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9t6mh" Dec 05 12:12:22 crc kubenswrapper[4807]: I1205 12:12:22.988169 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p67q6"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.617290 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-927f9"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.618219 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-927f9" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="registry-server" containerID="cri-o://30fa918558f678fccbbd25be63737e328608cb478732caffe141ed136dad9c0e" gracePeriod=30 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.623679 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4d4jf"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.624014 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4d4jf" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="registry-server" containerID="cri-o://9f915fda83de1bf39c93f25b297d80aa565698237cb2292c96519aabb44fbf9e" gracePeriod=30 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.638548 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bchkg"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.638754 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" containerID="cri-o://0fbf59fc0fd565cbd9ea441a4740f73c7e54bc81488acc206b0aad55003e8517" gracePeriod=30 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.642943 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsv4c"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.643186 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xsv4c" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="registry-server" containerID="cri-o://a20b6c3dbdd61806689482174168b15fd1a791798889448104f7fca8f9b6af83" gracePeriod=30 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.658142 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9c8wq"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.658501 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9c8wq" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="registry-server" containerID="cri-o://ec6d205c018c061bff0ca6c28357e23891d7f1a2f77b1eeb73790355d74cd903" gracePeriod=30 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.669418 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m26dt"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.671261 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.673601 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m26dt"] Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.782807 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/47587201-3528-4f92-8e07-09afeedcc8eb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.783225 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47587201-3528-4f92-8e07-09afeedcc8eb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.783266 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dbpt\" (UniqueName: \"kubernetes.io/projected/47587201-3528-4f92-8e07-09afeedcc8eb-kube-api-access-6dbpt\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.884473 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/47587201-3528-4f92-8e07-09afeedcc8eb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.884579 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47587201-3528-4f92-8e07-09afeedcc8eb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.884617 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dbpt\" (UniqueName: \"kubernetes.io/projected/47587201-3528-4f92-8e07-09afeedcc8eb-kube-api-access-6dbpt\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.885952 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47587201-3528-4f92-8e07-09afeedcc8eb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.891688 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/47587201-3528-4f92-8e07-09afeedcc8eb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.897449 4807 generic.go:334] "Generic (PLEG): container finished" podID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerID="30fa918558f678fccbbd25be63737e328608cb478732caffe141ed136dad9c0e" exitCode=0 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.897517 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-927f9" event={"ID":"55077615-de5e-48c6-9be0-2f3c8bbabae7","Type":"ContainerDied","Data":"30fa918558f678fccbbd25be63737e328608cb478732caffe141ed136dad9c0e"} Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.901884 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dbpt\" (UniqueName: \"kubernetes.io/projected/47587201-3528-4f92-8e07-09afeedcc8eb-kube-api-access-6dbpt\") pod \"marketplace-operator-79b997595-m26dt\" (UID: \"47587201-3528-4f92-8e07-09afeedcc8eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.902433 4807 generic.go:334] "Generic (PLEG): container finished" podID="2b219446-dca8-4aec-b300-92eb952d2a89" containerID="a20b6c3dbdd61806689482174168b15fd1a791798889448104f7fca8f9b6af83" exitCode=0 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.902489 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsv4c" event={"ID":"2b219446-dca8-4aec-b300-92eb952d2a89","Type":"ContainerDied","Data":"a20b6c3dbdd61806689482174168b15fd1a791798889448104f7fca8f9b6af83"} Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.904297 4807 generic.go:334] "Generic (PLEG): container finished" podID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerID="0fbf59fc0fd565cbd9ea441a4740f73c7e54bc81488acc206b0aad55003e8517" exitCode=0 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.904366 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" event={"ID":"71ce4423-09bb-43f1-a7a9-86143ea5dd9d","Type":"ContainerDied","Data":"0fbf59fc0fd565cbd9ea441a4740f73c7e54bc81488acc206b0aad55003e8517"} Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.904395 4807 scope.go:117] "RemoveContainer" containerID="f4ab674188cc0f8873a0882d7a969dc85235e8c5d9409617ec7dfdda9c4a1b16" Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.907269 4807 generic.go:334] "Generic (PLEG): container finished" podID="572bedd6-2920-481b-a356-c96fa8766456" containerID="ec6d205c018c061bff0ca6c28357e23891d7f1a2f77b1eeb73790355d74cd903" exitCode=0 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.907329 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9c8wq" event={"ID":"572bedd6-2920-481b-a356-c96fa8766456","Type":"ContainerDied","Data":"ec6d205c018c061bff0ca6c28357e23891d7f1a2f77b1eeb73790355d74cd903"} Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.909451 4807 generic.go:334] "Generic (PLEG): container finished" podID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerID="9f915fda83de1bf39c93f25b297d80aa565698237cb2292c96519aabb44fbf9e" exitCode=0 Dec 05 12:12:33 crc kubenswrapper[4807]: I1205 12:12:33.909480 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d4jf" event={"ID":"c9af52ad-e3d9-426f-9adf-a9880437d83f","Type":"ContainerDied","Data":"9f915fda83de1bf39c93f25b297d80aa565698237cb2292c96519aabb44fbf9e"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.021888 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.106424 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.288408 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-catalog-content\") pod \"55077615-de5e-48c6-9be0-2f3c8bbabae7\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.288562 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t4gl\" (UniqueName: \"kubernetes.io/projected/55077615-de5e-48c6-9be0-2f3c8bbabae7-kube-api-access-5t4gl\") pod \"55077615-de5e-48c6-9be0-2f3c8bbabae7\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.288617 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-utilities\") pod \"55077615-de5e-48c6-9be0-2f3c8bbabae7\" (UID: \"55077615-de5e-48c6-9be0-2f3c8bbabae7\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.289838 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-utilities" (OuterVolumeSpecName: "utilities") pod "55077615-de5e-48c6-9be0-2f3c8bbabae7" (UID: "55077615-de5e-48c6-9be0-2f3c8bbabae7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.295149 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55077615-de5e-48c6-9be0-2f3c8bbabae7-kube-api-access-5t4gl" (OuterVolumeSpecName: "kube-api-access-5t4gl") pod "55077615-de5e-48c6-9be0-2f3c8bbabae7" (UID: "55077615-de5e-48c6-9be0-2f3c8bbabae7"). InnerVolumeSpecName "kube-api-access-5t4gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.341137 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.345827 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.353098 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.365721 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55077615-de5e-48c6-9be0-2f3c8bbabae7" (UID: "55077615-de5e-48c6-9be0-2f3c8bbabae7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.368395 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392295 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f72wv\" (UniqueName: \"kubernetes.io/projected/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-kube-api-access-f72wv\") pod \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392338 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5z6f\" (UniqueName: \"kubernetes.io/projected/572bedd6-2920-481b-a356-c96fa8766456-kube-api-access-j5z6f\") pod \"572bedd6-2920-481b-a356-c96fa8766456\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392362 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-operator-metrics\") pod \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392390 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-utilities\") pod \"2b219446-dca8-4aec-b300-92eb952d2a89\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392413 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-catalog-content\") pod \"2b219446-dca8-4aec-b300-92eb952d2a89\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392433 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-catalog-content\") pod \"572bedd6-2920-481b-a356-c96fa8766456\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392462 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-catalog-content\") pod \"c9af52ad-e3d9-426f-9adf-a9880437d83f\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392484 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfj9l\" (UniqueName: \"kubernetes.io/projected/c9af52ad-e3d9-426f-9adf-a9880437d83f-kube-api-access-zfj9l\") pod \"c9af52ad-e3d9-426f-9adf-a9880437d83f\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392510 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-utilities\") pod \"572bedd6-2920-481b-a356-c96fa8766456\" (UID: \"572bedd6-2920-481b-a356-c96fa8766456\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.392543 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-trusted-ca\") pod \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\" (UID: \"71ce4423-09bb-43f1-a7a9-86143ea5dd9d\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.395920 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-utilities" (OuterVolumeSpecName: "utilities") pod "572bedd6-2920-481b-a356-c96fa8766456" (UID: "572bedd6-2920-481b-a356-c96fa8766456"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.396380 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-utilities" (OuterVolumeSpecName: "utilities") pod "2b219446-dca8-4aec-b300-92eb952d2a89" (UID: "2b219446-dca8-4aec-b300-92eb952d2a89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.396776 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "71ce4423-09bb-43f1-a7a9-86143ea5dd9d" (UID: "71ce4423-09bb-43f1-a7a9-86143ea5dd9d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.397305 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-kube-api-access-f72wv" (OuterVolumeSpecName: "kube-api-access-f72wv") pod "71ce4423-09bb-43f1-a7a9-86143ea5dd9d" (UID: "71ce4423-09bb-43f1-a7a9-86143ea5dd9d"). InnerVolumeSpecName "kube-api-access-f72wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.397484 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572bedd6-2920-481b-a356-c96fa8766456-kube-api-access-j5z6f" (OuterVolumeSpecName: "kube-api-access-j5z6f") pod "572bedd6-2920-481b-a356-c96fa8766456" (UID: "572bedd6-2920-481b-a356-c96fa8766456"). InnerVolumeSpecName "kube-api-access-j5z6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.397720 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9af52ad-e3d9-426f-9adf-a9880437d83f-kube-api-access-zfj9l" (OuterVolumeSpecName: "kube-api-access-zfj9l") pod "c9af52ad-e3d9-426f-9adf-a9880437d83f" (UID: "c9af52ad-e3d9-426f-9adf-a9880437d83f"). InnerVolumeSpecName "kube-api-access-zfj9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.398184 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "71ce4423-09bb-43f1-a7a9-86143ea5dd9d" (UID: "71ce4423-09bb-43f1-a7a9-86143ea5dd9d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406157 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5z6f\" (UniqueName: \"kubernetes.io/projected/572bedd6-2920-481b-a356-c96fa8766456-kube-api-access-j5z6f\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406190 4807 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406243 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406257 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t4gl\" (UniqueName: \"kubernetes.io/projected/55077615-de5e-48c6-9be0-2f3c8bbabae7-kube-api-access-5t4gl\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406270 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfj9l\" (UniqueName: \"kubernetes.io/projected/c9af52ad-e3d9-426f-9adf-a9880437d83f-kube-api-access-zfj9l\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406282 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406294 4807 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406307 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406318 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55077615-de5e-48c6-9be0-2f3c8bbabae7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.406330 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f72wv\" (UniqueName: \"kubernetes.io/projected/71ce4423-09bb-43f1-a7a9-86143ea5dd9d-kube-api-access-f72wv\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.420572 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b219446-dca8-4aec-b300-92eb952d2a89" (UID: "2b219446-dca8-4aec-b300-92eb952d2a89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.452512 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9af52ad-e3d9-426f-9adf-a9880437d83f" (UID: "c9af52ad-e3d9-426f-9adf-a9880437d83f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.507372 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm6jv\" (UniqueName: \"kubernetes.io/projected/2b219446-dca8-4aec-b300-92eb952d2a89-kube-api-access-lm6jv\") pod \"2b219446-dca8-4aec-b300-92eb952d2a89\" (UID: \"2b219446-dca8-4aec-b300-92eb952d2a89\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.507440 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-utilities\") pod \"c9af52ad-e3d9-426f-9adf-a9880437d83f\" (UID: \"c9af52ad-e3d9-426f-9adf-a9880437d83f\") " Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.508193 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.508213 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b219446-dca8-4aec-b300-92eb952d2a89-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.509058 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-utilities" (OuterVolumeSpecName: "utilities") pod "c9af52ad-e3d9-426f-9adf-a9880437d83f" (UID: "c9af52ad-e3d9-426f-9adf-a9880437d83f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.511886 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b219446-dca8-4aec-b300-92eb952d2a89-kube-api-access-lm6jv" (OuterVolumeSpecName: "kube-api-access-lm6jv") pod "2b219446-dca8-4aec-b300-92eb952d2a89" (UID: "2b219446-dca8-4aec-b300-92eb952d2a89"). InnerVolumeSpecName "kube-api-access-lm6jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.521216 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "572bedd6-2920-481b-a356-c96fa8766456" (UID: "572bedd6-2920-481b-a356-c96fa8766456"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.610099 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm6jv\" (UniqueName: \"kubernetes.io/projected/2b219446-dca8-4aec-b300-92eb952d2a89-kube-api-access-lm6jv\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.610171 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9af52ad-e3d9-426f-9adf-a9880437d83f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.610192 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572bedd6-2920-481b-a356-c96fa8766456-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.652804 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m26dt"] Dec 05 12:12:34 crc kubenswrapper[4807]: W1205 12:12:34.665421 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47587201_3528_4f92_8e07_09afeedcc8eb.slice/crio-846530e1e3be21aecd535e93e0af10853446060096e4a049d83ba1c91d3cc4a5 WatchSource:0}: Error finding container 846530e1e3be21aecd535e93e0af10853446060096e4a049d83ba1c91d3cc4a5: Status 404 returned error can't find the container with id 846530e1e3be21aecd535e93e0af10853446060096e4a049d83ba1c91d3cc4a5 Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.915951 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" event={"ID":"71ce4423-09bb-43f1-a7a9-86143ea5dd9d","Type":"ContainerDied","Data":"23c1bf196b7eed62bc7bfe71c73f9ee97306930633c6551bf853a8117aab78e4"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.916261 4807 scope.go:117] "RemoveContainer" containerID="0fbf59fc0fd565cbd9ea441a4740f73c7e54bc81488acc206b0aad55003e8517" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.916024 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bchkg" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.920177 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9c8wq" event={"ID":"572bedd6-2920-481b-a356-c96fa8766456","Type":"ContainerDied","Data":"221acccf132fea0e0ef368c20532cfaced3dcb9d04b9ef67bf58d4a8bcd4fa79"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.920263 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9c8wq" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.924589 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d4jf" event={"ID":"c9af52ad-e3d9-426f-9adf-a9880437d83f","Type":"ContainerDied","Data":"89b2e81423724508e8588cb7b9a6816cf33c9fcdf80b64ea6b7f6a97a45fa081"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.924676 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d4jf" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.931675 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-927f9" event={"ID":"55077615-de5e-48c6-9be0-2f3c8bbabae7","Type":"ContainerDied","Data":"5204b0542fac4d880f8a962429d4e26e08dc5e21f10268badb9ecb55e1d8b905"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.931699 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-927f9" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.933469 4807 scope.go:117] "RemoveContainer" containerID="ec6d205c018c061bff0ca6c28357e23891d7f1a2f77b1eeb73790355d74cd903" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.933767 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsv4c" event={"ID":"2b219446-dca8-4aec-b300-92eb952d2a89","Type":"ContainerDied","Data":"62ebfc69a567e5f67fbb0a522668836e339b37319b4964363fa0d649284b5efc"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.933813 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsv4c" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.935317 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" event={"ID":"47587201-3528-4f92-8e07-09afeedcc8eb","Type":"ContainerStarted","Data":"aae61fb3900c7015885dd4cc1e73aed1f7ed10709b8d0010486e880496e03196"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.935354 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" event={"ID":"47587201-3528-4f92-8e07-09afeedcc8eb","Type":"ContainerStarted","Data":"846530e1e3be21aecd535e93e0af10853446060096e4a049d83ba1c91d3cc4a5"} Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.937462 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.937580 4807 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-m26dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.937648 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" podUID="47587201-3528-4f92-8e07-09afeedcc8eb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.959466 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" podStartSLOduration=1.959427536 podStartE2EDuration="1.959427536s" podCreationTimestamp="2025-12-05 12:12:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:12:34.956990062 +0000 UTC m=+384.450853331" watchObservedRunningTime="2025-12-05 12:12:34.959427536 +0000 UTC m=+384.453290805" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.961247 4807 scope.go:117] "RemoveContainer" containerID="53552f960a25955ee166a8fc5f5e35dbfe1fcdf5bb616262a758ff44b97a5476" Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.973316 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bchkg"] Dec 05 12:12:34 crc kubenswrapper[4807]: I1205 12:12:34.998601 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bchkg"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.000166 4807 scope.go:117] "RemoveContainer" containerID="edea6372167268d889da61726b7c132644f31108c85bb3ffcfdff2012e895c74" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.005052 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-927f9"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.021901 4807 scope.go:117] "RemoveContainer" containerID="9f915fda83de1bf39c93f25b297d80aa565698237cb2292c96519aabb44fbf9e" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.025287 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-927f9"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.036148 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4d4jf"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.043836 4807 scope.go:117] "RemoveContainer" containerID="1b67e88b11c6a43a8a0a2bfc78c6f1fa45d8b9257497605bd462fd8d52619040" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.048768 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4d4jf"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.052774 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9c8wq"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.057499 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9c8wq"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.060173 4807 scope.go:117] "RemoveContainer" containerID="11d02f984ff2b89023f221c6c2e2289932914bd31466232fd48a8d8ee7ead036" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.062888 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsv4c"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.067249 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsv4c"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.073876 4807 scope.go:117] "RemoveContainer" containerID="30fa918558f678fccbbd25be63737e328608cb478732caffe141ed136dad9c0e" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.092833 4807 scope.go:117] "RemoveContainer" containerID="16ca63af1a3d6226cddced2d618ee976bad5b83bb3afdd6cd65e7e1e392019d4" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.107275 4807 scope.go:117] "RemoveContainer" containerID="df498dc0e65691752278b2aee36eb3c68a49e7a6a273a8780fc32a07461f6cd6" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.140436 4807 scope.go:117] "RemoveContainer" containerID="a20b6c3dbdd61806689482174168b15fd1a791798889448104f7fca8f9b6af83" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.154638 4807 scope.go:117] "RemoveContainer" containerID="4c204904a8921e521511678e7272e3a091e1052f464bfc06dd93da1afe10a9c3" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.165752 4807 scope.go:117] "RemoveContainer" containerID="60cf6d9ebeb17ca53a6db4cc25720ffc7dc81ba5f78a2be8790a4598d6a158a7" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.243823 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" path="/var/lib/kubelet/pods/2b219446-dca8-4aec-b300-92eb952d2a89/volumes" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.244588 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" path="/var/lib/kubelet/pods/55077615-de5e-48c6-9be0-2f3c8bbabae7/volumes" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.245327 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572bedd6-2920-481b-a356-c96fa8766456" path="/var/lib/kubelet/pods/572bedd6-2920-481b-a356-c96fa8766456/volumes" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.247441 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" path="/var/lib/kubelet/pods/71ce4423-09bb-43f1-a7a9-86143ea5dd9d/volumes" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.251262 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" path="/var/lib/kubelet/pods/c9af52ad-e3d9-426f-9adf-a9880437d83f/volumes" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.643945 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6f2l5"] Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644234 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644253 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644268 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644279 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644296 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644310 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644324 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644335 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644353 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644364 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644379 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644389 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644403 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644414 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644429 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644438 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644455 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644465 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644477 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644488 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644502 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644512 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="extract-utilities" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644546 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644556 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644571 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644580 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: E1205 12:12:35.644594 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644604 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="extract-content" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644744 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644760 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b219446-dca8-4aec-b300-92eb952d2a89" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644776 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="71ce4423-09bb-43f1-a7a9-86143ea5dd9d" containerName="marketplace-operator" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644795 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="55077615-de5e-48c6-9be0-2f3c8bbabae7" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644811 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="572bedd6-2920-481b-a356-c96fa8766456" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.644822 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9af52ad-e3d9-426f-9adf-a9880437d83f" containerName="registry-server" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.646011 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.654507 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.661389 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6f2l5"] Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.730142 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c28504-1b38-4615-98f2-529b0bfb088f-utilities\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.730206 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c28504-1b38-4615-98f2-529b0bfb088f-catalog-content\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.730290 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjpmh\" (UniqueName: \"kubernetes.io/projected/93c28504-1b38-4615-98f2-529b0bfb088f-kube-api-access-cjpmh\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.831251 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjpmh\" (UniqueName: \"kubernetes.io/projected/93c28504-1b38-4615-98f2-529b0bfb088f-kube-api-access-cjpmh\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.831998 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c28504-1b38-4615-98f2-529b0bfb088f-utilities\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.832428 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c28504-1b38-4615-98f2-529b0bfb088f-utilities\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.832978 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c28504-1b38-4615-98f2-529b0bfb088f-catalog-content\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.832509 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c28504-1b38-4615-98f2-529b0bfb088f-catalog-content\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.855113 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjpmh\" (UniqueName: \"kubernetes.io/projected/93c28504-1b38-4615-98f2-529b0bfb088f-kube-api-access-cjpmh\") pod \"redhat-marketplace-6f2l5\" (UID: \"93c28504-1b38-4615-98f2-529b0bfb088f\") " pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.947414 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-m26dt" Dec 05 12:12:35 crc kubenswrapper[4807]: I1205 12:12:35.966224 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.242757 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hmcwq"] Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.247891 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.253747 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.255887 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmcwq"] Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.339629 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bd9595-697e-4315-a23f-11f90b8a8104-utilities\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.339733 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bd9595-697e-4315-a23f-11f90b8a8104-catalog-content\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.341302 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w68fv\" (UniqueName: \"kubernetes.io/projected/c5bd9595-697e-4315-a23f-11f90b8a8104-kube-api-access-w68fv\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.366662 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6f2l5"] Dec 05 12:12:36 crc kubenswrapper[4807]: W1205 12:12:36.375676 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93c28504_1b38_4615_98f2_529b0bfb088f.slice/crio-efb7b338c0e58e1da200f77f50423b217dd5a0cabd1d23c0423f69749dc6fd8d WatchSource:0}: Error finding container efb7b338c0e58e1da200f77f50423b217dd5a0cabd1d23c0423f69749dc6fd8d: Status 404 returned error can't find the container with id efb7b338c0e58e1da200f77f50423b217dd5a0cabd1d23c0423f69749dc6fd8d Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.442893 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bd9595-697e-4315-a23f-11f90b8a8104-catalog-content\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.442996 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w68fv\" (UniqueName: \"kubernetes.io/projected/c5bd9595-697e-4315-a23f-11f90b8a8104-kube-api-access-w68fv\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.443053 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bd9595-697e-4315-a23f-11f90b8a8104-utilities\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.443555 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5bd9595-697e-4315-a23f-11f90b8a8104-utilities\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.443807 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5bd9595-697e-4315-a23f-11f90b8a8104-catalog-content\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.466137 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w68fv\" (UniqueName: \"kubernetes.io/projected/c5bd9595-697e-4315-a23f-11f90b8a8104-kube-api-access-w68fv\") pod \"redhat-operators-hmcwq\" (UID: \"c5bd9595-697e-4315-a23f-11f90b8a8104\") " pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.574720 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.952957 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmcwq"] Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.954599 4807 generic.go:334] "Generic (PLEG): container finished" podID="93c28504-1b38-4615-98f2-529b0bfb088f" containerID="22ccaab0ccaec43fb5a00fecad17b3cad54ca0310d156734612d02f9400e647b" exitCode=0 Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.954683 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6f2l5" event={"ID":"93c28504-1b38-4615-98f2-529b0bfb088f","Type":"ContainerDied","Data":"22ccaab0ccaec43fb5a00fecad17b3cad54ca0310d156734612d02f9400e647b"} Dec 05 12:12:36 crc kubenswrapper[4807]: I1205 12:12:36.954720 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6f2l5" event={"ID":"93c28504-1b38-4615-98f2-529b0bfb088f","Type":"ContainerStarted","Data":"efb7b338c0e58e1da200f77f50423b217dd5a0cabd1d23c0423f69749dc6fd8d"} Dec 05 12:12:36 crc kubenswrapper[4807]: W1205 12:12:36.960698 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5bd9595_697e_4315_a23f_11f90b8a8104.slice/crio-0ebb64b609390ed986798c6a08104dd15f6b81e8452ea07c82d833652401053d WatchSource:0}: Error finding container 0ebb64b609390ed986798c6a08104dd15f6b81e8452ea07c82d833652401053d: Status 404 returned error can't find the container with id 0ebb64b609390ed986798c6a08104dd15f6b81e8452ea07c82d833652401053d Dec 05 12:12:37 crc kubenswrapper[4807]: I1205 12:12:37.962415 4807 generic.go:334] "Generic (PLEG): container finished" podID="c5bd9595-697e-4315-a23f-11f90b8a8104" containerID="9eb913da437a2b942c11bdfad04e62487d802c84b5fc7dfd7b8c4109cde591d6" exitCode=0 Dec 05 12:12:37 crc kubenswrapper[4807]: I1205 12:12:37.962638 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmcwq" event={"ID":"c5bd9595-697e-4315-a23f-11f90b8a8104","Type":"ContainerDied","Data":"9eb913da437a2b942c11bdfad04e62487d802c84b5fc7dfd7b8c4109cde591d6"} Dec 05 12:12:37 crc kubenswrapper[4807]: I1205 12:12:37.963030 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmcwq" event={"ID":"c5bd9595-697e-4315-a23f-11f90b8a8104","Type":"ContainerStarted","Data":"0ebb64b609390ed986798c6a08104dd15f6b81e8452ea07c82d833652401053d"} Dec 05 12:12:37 crc kubenswrapper[4807]: I1205 12:12:37.967811 4807 generic.go:334] "Generic (PLEG): container finished" podID="93c28504-1b38-4615-98f2-529b0bfb088f" containerID="c762c3083cafd9f7ceb2a0152e9ea67ea609e740a2f7d0098fc5dbdf7cda9be9" exitCode=0 Dec 05 12:12:37 crc kubenswrapper[4807]: I1205 12:12:37.968116 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6f2l5" event={"ID":"93c28504-1b38-4615-98f2-529b0bfb088f","Type":"ContainerDied","Data":"c762c3083cafd9f7ceb2a0152e9ea67ea609e740a2f7d0098fc5dbdf7cda9be9"} Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.044425 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-84xwf"] Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.045514 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.049114 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.053672 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84xwf"] Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.061230 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc5n5\" (UniqueName: \"kubernetes.io/projected/aa7d7861-9998-4174-974a-47b939d4c4bf-kube-api-access-zc5n5\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.061350 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7d7861-9998-4174-974a-47b939d4c4bf-utilities\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.061412 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7d7861-9998-4174-974a-47b939d4c4bf-catalog-content\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.162734 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc5n5\" (UniqueName: \"kubernetes.io/projected/aa7d7861-9998-4174-974a-47b939d4c4bf-kube-api-access-zc5n5\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.162800 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7d7861-9998-4174-974a-47b939d4c4bf-utilities\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.162826 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7d7861-9998-4174-974a-47b939d4c4bf-catalog-content\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.163262 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7d7861-9998-4174-974a-47b939d4c4bf-catalog-content\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.163379 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7d7861-9998-4174-974a-47b939d4c4bf-utilities\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.181469 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc5n5\" (UniqueName: \"kubernetes.io/projected/aa7d7861-9998-4174-974a-47b939d4c4bf-kube-api-access-zc5n5\") pod \"certified-operators-84xwf\" (UID: \"aa7d7861-9998-4174-974a-47b939d4c4bf\") " pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.372316 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.649344 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v99qd"] Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.650603 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.652618 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.666653 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v99qd"] Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.667487 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd4r2\" (UniqueName: \"kubernetes.io/projected/2b7c31e6-006a-450f-af53-e07c2fc71499-kube-api-access-bd4r2\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.667552 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7c31e6-006a-450f-af53-e07c2fc71499-catalog-content\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.667627 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7c31e6-006a-450f-af53-e07c2fc71499-utilities\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.768496 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7c31e6-006a-450f-af53-e07c2fc71499-utilities\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.768590 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd4r2\" (UniqueName: \"kubernetes.io/projected/2b7c31e6-006a-450f-af53-e07c2fc71499-kube-api-access-bd4r2\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.768625 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7c31e6-006a-450f-af53-e07c2fc71499-catalog-content\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.769016 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7c31e6-006a-450f-af53-e07c2fc71499-utilities\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.769029 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7c31e6-006a-450f-af53-e07c2fc71499-catalog-content\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.792439 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84xwf"] Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.794620 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd4r2\" (UniqueName: \"kubernetes.io/projected/2b7c31e6-006a-450f-af53-e07c2fc71499-kube-api-access-bd4r2\") pod \"community-operators-v99qd\" (UID: \"2b7c31e6-006a-450f-af53-e07c2fc71499\") " pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.974813 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmcwq" event={"ID":"c5bd9595-697e-4315-a23f-11f90b8a8104","Type":"ContainerStarted","Data":"12a033422da324b5c2328ddb749b569b00b4b7a2e6c03c90356942e47c60d098"} Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.977101 4807 generic.go:334] "Generic (PLEG): container finished" podID="aa7d7861-9998-4174-974a-47b939d4c4bf" containerID="fcfd52b203aeed144912a068c2a8eba5426db3fc7fd1a396d2a69240418b2ae3" exitCode=0 Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.977170 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84xwf" event={"ID":"aa7d7861-9998-4174-974a-47b939d4c4bf","Type":"ContainerDied","Data":"fcfd52b203aeed144912a068c2a8eba5426db3fc7fd1a396d2a69240418b2ae3"} Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.977195 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84xwf" event={"ID":"aa7d7861-9998-4174-974a-47b939d4c4bf","Type":"ContainerStarted","Data":"fa8af4d4ccd4bbb818217ae749d003fbbb4eb0d3320773536aa0b4a186c58957"} Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.979189 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:38 crc kubenswrapper[4807]: I1205 12:12:38.979563 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6f2l5" event={"ID":"93c28504-1b38-4615-98f2-529b0bfb088f","Type":"ContainerStarted","Data":"6f2163b3a469684488a8bbb797955dced3b2d331efe406dd0d130f83fac4d087"} Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.031332 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6f2l5" podStartSLOduration=2.652228959 podStartE2EDuration="4.031313774s" podCreationTimestamp="2025-12-05 12:12:35 +0000 UTC" firstStartedPulling="2025-12-05 12:12:36.960811527 +0000 UTC m=+386.454674796" lastFinishedPulling="2025-12-05 12:12:38.339896342 +0000 UTC m=+387.833759611" observedRunningTime="2025-12-05 12:12:39.030262916 +0000 UTC m=+388.524126195" watchObservedRunningTime="2025-12-05 12:12:39.031313774 +0000 UTC m=+388.525177043" Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.378399 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v99qd"] Dec 05 12:12:39 crc kubenswrapper[4807]: W1205 12:12:39.386457 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b7c31e6_006a_450f_af53_e07c2fc71499.slice/crio-bd53a0ad69f612c3a63cc7eae986d94e5478f7eaf2fddddfbfb5fc38d23733c2 WatchSource:0}: Error finding container bd53a0ad69f612c3a63cc7eae986d94e5478f7eaf2fddddfbfb5fc38d23733c2: Status 404 returned error can't find the container with id bd53a0ad69f612c3a63cc7eae986d94e5478f7eaf2fddddfbfb5fc38d23733c2 Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.986848 4807 generic.go:334] "Generic (PLEG): container finished" podID="c5bd9595-697e-4315-a23f-11f90b8a8104" containerID="12a033422da324b5c2328ddb749b569b00b4b7a2e6c03c90356942e47c60d098" exitCode=0 Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.986950 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmcwq" event={"ID":"c5bd9595-697e-4315-a23f-11f90b8a8104","Type":"ContainerDied","Data":"12a033422da324b5c2328ddb749b569b00b4b7a2e6c03c90356942e47c60d098"} Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.989073 4807 generic.go:334] "Generic (PLEG): container finished" podID="aa7d7861-9998-4174-974a-47b939d4c4bf" containerID="354867dc95346d1bec637409660382c9c2a06543d8aaa0606a0e6bd9eef1e3c5" exitCode=0 Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.989149 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84xwf" event={"ID":"aa7d7861-9998-4174-974a-47b939d4c4bf","Type":"ContainerDied","Data":"354867dc95346d1bec637409660382c9c2a06543d8aaa0606a0e6bd9eef1e3c5"} Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.992122 4807 generic.go:334] "Generic (PLEG): container finished" podID="2b7c31e6-006a-450f-af53-e07c2fc71499" containerID="aed2a6e686eab1d5ed99f821349473b7951fb7052da0143be632f3c793e8216d" exitCode=0 Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.992202 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v99qd" event={"ID":"2b7c31e6-006a-450f-af53-e07c2fc71499","Type":"ContainerDied","Data":"aed2a6e686eab1d5ed99f821349473b7951fb7052da0143be632f3c793e8216d"} Dec 05 12:12:39 crc kubenswrapper[4807]: I1205 12:12:39.992221 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v99qd" event={"ID":"2b7c31e6-006a-450f-af53-e07c2fc71499","Type":"ContainerStarted","Data":"bd53a0ad69f612c3a63cc7eae986d94e5478f7eaf2fddddfbfb5fc38d23733c2"} Dec 05 12:12:41 crc kubenswrapper[4807]: I1205 12:12:41.002224 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v99qd" event={"ID":"2b7c31e6-006a-450f-af53-e07c2fc71499","Type":"ContainerStarted","Data":"22bab7dacc99068e3cd4df40036d6d67510932194292c9776096afc879149e24"} Dec 05 12:12:41 crc kubenswrapper[4807]: I1205 12:12:41.004679 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmcwq" event={"ID":"c5bd9595-697e-4315-a23f-11f90b8a8104","Type":"ContainerStarted","Data":"e108f14388e4d0f6690f1c012354a39940b06b6378c997babc2dc2d3d189fead"} Dec 05 12:12:41 crc kubenswrapper[4807]: I1205 12:12:41.006517 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84xwf" event={"ID":"aa7d7861-9998-4174-974a-47b939d4c4bf","Type":"ContainerStarted","Data":"40f44fb6fafc5f6ddcee894f6498c8bf1c9dd0793e6f8c757f9b97c27e7f8d55"} Dec 05 12:12:41 crc kubenswrapper[4807]: I1205 12:12:41.038207 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-84xwf" podStartSLOduration=1.418493706 podStartE2EDuration="3.03818807s" podCreationTimestamp="2025-12-05 12:12:38 +0000 UTC" firstStartedPulling="2025-12-05 12:12:38.978673792 +0000 UTC m=+388.472537061" lastFinishedPulling="2025-12-05 12:12:40.598368156 +0000 UTC m=+390.092231425" observedRunningTime="2025-12-05 12:12:41.0370549 +0000 UTC m=+390.530918189" watchObservedRunningTime="2025-12-05 12:12:41.03818807 +0000 UTC m=+390.532051339" Dec 05 12:12:41 crc kubenswrapper[4807]: I1205 12:12:41.058270 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hmcwq" podStartSLOduration=2.397883605 podStartE2EDuration="5.058246896s" podCreationTimestamp="2025-12-05 12:12:36 +0000 UTC" firstStartedPulling="2025-12-05 12:12:37.963901976 +0000 UTC m=+387.457765245" lastFinishedPulling="2025-12-05 12:12:40.624265267 +0000 UTC m=+390.118128536" observedRunningTime="2025-12-05 12:12:41.055166346 +0000 UTC m=+390.549029615" watchObservedRunningTime="2025-12-05 12:12:41.058246896 +0000 UTC m=+390.552110165" Dec 05 12:12:42 crc kubenswrapper[4807]: I1205 12:12:42.013173 4807 generic.go:334] "Generic (PLEG): container finished" podID="2b7c31e6-006a-450f-af53-e07c2fc71499" containerID="22bab7dacc99068e3cd4df40036d6d67510932194292c9776096afc879149e24" exitCode=0 Dec 05 12:12:42 crc kubenswrapper[4807]: I1205 12:12:42.014449 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v99qd" event={"ID":"2b7c31e6-006a-450f-af53-e07c2fc71499","Type":"ContainerDied","Data":"22bab7dacc99068e3cd4df40036d6d67510932194292c9776096afc879149e24"} Dec 05 12:12:44 crc kubenswrapper[4807]: I1205 12:12:44.036632 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v99qd" event={"ID":"2b7c31e6-006a-450f-af53-e07c2fc71499","Type":"ContainerStarted","Data":"d750ec1581a6cbc103d0072bc772ed9504e83bd3192e6103de7e6ad58695dac4"} Dec 05 12:12:44 crc kubenswrapper[4807]: I1205 12:12:44.056490 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v99qd" podStartSLOduration=3.653587522 podStartE2EDuration="6.056471342s" podCreationTimestamp="2025-12-05 12:12:38 +0000 UTC" firstStartedPulling="2025-12-05 12:12:39.993246071 +0000 UTC m=+389.487109340" lastFinishedPulling="2025-12-05 12:12:42.396129891 +0000 UTC m=+391.889993160" observedRunningTime="2025-12-05 12:12:44.055685646 +0000 UTC m=+393.549548915" watchObservedRunningTime="2025-12-05 12:12:44.056471342 +0000 UTC m=+393.550334611" Dec 05 12:12:45 crc kubenswrapper[4807]: I1205 12:12:45.967746 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:45 crc kubenswrapper[4807]: I1205 12:12:45.968583 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:46 crc kubenswrapper[4807]: I1205 12:12:46.010513 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:46 crc kubenswrapper[4807]: I1205 12:12:46.092197 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6f2l5" Dec 05 12:12:46 crc kubenswrapper[4807]: I1205 12:12:46.575521 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:46 crc kubenswrapper[4807]: I1205 12:12:46.575588 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:46 crc kubenswrapper[4807]: I1205 12:12:46.622797 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:47 crc kubenswrapper[4807]: I1205 12:12:47.105682 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hmcwq" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.029503 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" podUID="2ea71070-7150-4f2d-b531-08f22697c479" containerName="registry" containerID="cri-o://4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400" gracePeriod=30 Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.372547 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.372919 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.417279 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.446241 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.501926 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-bound-sa-token\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.502062 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-trusted-ca\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.502095 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ea71070-7150-4f2d-b531-08f22697c479-installation-pull-secrets\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.502142 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-registry-tls\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.502323 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.502383 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ea71070-7150-4f2d-b531-08f22697c479-ca-trust-extracted\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.502419 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqpgf\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-kube-api-access-bqpgf\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.502457 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-registry-certificates\") pod \"2ea71070-7150-4f2d-b531-08f22697c479\" (UID: \"2ea71070-7150-4f2d-b531-08f22697c479\") " Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.503003 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.503401 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.510929 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.511015 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea71070-7150-4f2d-b531-08f22697c479-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.512563 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-kube-api-access-bqpgf" (OuterVolumeSpecName: "kube-api-access-bqpgf") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "kube-api-access-bqpgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.512832 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.516244 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.520816 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ea71070-7150-4f2d-b531-08f22697c479-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2ea71070-7150-4f2d-b531-08f22697c479" (UID: "2ea71070-7150-4f2d-b531-08f22697c479"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.603876 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.603943 4807 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ea71070-7150-4f2d-b531-08f22697c479-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.603958 4807 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.603967 4807 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ea71070-7150-4f2d-b531-08f22697c479-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.603979 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqpgf\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-kube-api-access-bqpgf\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.603988 4807 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ea71070-7150-4f2d-b531-08f22697c479-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.603998 4807 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ea71070-7150-4f2d-b531-08f22697c479-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.980246 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:48 crc kubenswrapper[4807]: I1205 12:12:48.980315 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.025688 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.070754 4807 generic.go:334] "Generic (PLEG): container finished" podID="2ea71070-7150-4f2d-b531-08f22697c479" containerID="4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400" exitCode=0 Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.070809 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" event={"ID":"2ea71070-7150-4f2d-b531-08f22697c479","Type":"ContainerDied","Data":"4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400"} Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.070860 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" event={"ID":"2ea71070-7150-4f2d-b531-08f22697c479","Type":"ContainerDied","Data":"c092fcb75d29fd93436d1cf2531daa73c2d913de0c78d7b5f3123efc7581c3c8"} Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.070883 4807 scope.go:117] "RemoveContainer" containerID="4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.070893 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p67q6" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.088816 4807 scope.go:117] "RemoveContainer" containerID="4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400" Dec 05 12:12:49 crc kubenswrapper[4807]: E1205 12:12:49.089230 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400\": container with ID starting with 4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400 not found: ID does not exist" containerID="4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.089259 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400"} err="failed to get container status \"4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400\": rpc error: code = NotFound desc = could not find container \"4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400\": container with ID starting with 4c6348a05627f6a3fc88e23d2986e9ed2b31df0671dc64bf1c388c050d9fe400 not found: ID does not exist" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.107204 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p67q6"] Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.122319 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p67q6"] Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.124360 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v99qd" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.134841 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-84xwf" Dec 05 12:12:49 crc kubenswrapper[4807]: I1205 12:12:49.243769 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea71070-7150-4f2d-b531-08f22697c479" path="/var/lib/kubelet/pods/2ea71070-7150-4f2d-b531-08f22697c479/volumes" Dec 05 12:12:52 crc kubenswrapper[4807]: I1205 12:12:52.466200 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:12:52 crc kubenswrapper[4807]: I1205 12:12:52.466261 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:13:22 crc kubenswrapper[4807]: I1205 12:13:22.467037 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:13:22 crc kubenswrapper[4807]: I1205 12:13:22.467654 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:13:22 crc kubenswrapper[4807]: I1205 12:13:22.467702 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:13:22 crc kubenswrapper[4807]: I1205 12:13:22.468313 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61fbdac8b8e46032c2f8db25d5c108f1020da47f0c78713ac4dad89d5fdc1819"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:13:22 crc kubenswrapper[4807]: I1205 12:13:22.468367 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://61fbdac8b8e46032c2f8db25d5c108f1020da47f0c78713ac4dad89d5fdc1819" gracePeriod=600 Dec 05 12:13:23 crc kubenswrapper[4807]: I1205 12:13:23.251073 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="61fbdac8b8e46032c2f8db25d5c108f1020da47f0c78713ac4dad89d5fdc1819" exitCode=0 Dec 05 12:13:23 crc kubenswrapper[4807]: I1205 12:13:23.251151 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"61fbdac8b8e46032c2f8db25d5c108f1020da47f0c78713ac4dad89d5fdc1819"} Dec 05 12:13:23 crc kubenswrapper[4807]: I1205 12:13:23.251671 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"d4a9bf0256da38a66ac347b1c3d3069dc78bcbfdcd0d597ffe57a7702a56bd70"} Dec 05 12:13:23 crc kubenswrapper[4807]: I1205 12:13:23.251699 4807 scope.go:117] "RemoveContainer" containerID="b2f14fd81fffefe02b3a631df1d77f62b11304f99e869eb2132f19cdd83b855a" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.164320 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx"] Dec 05 12:15:00 crc kubenswrapper[4807]: E1205 12:15:00.165670 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea71070-7150-4f2d-b531-08f22697c479" containerName="registry" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.165739 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea71070-7150-4f2d-b531-08f22697c479" containerName="registry" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.166364 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea71070-7150-4f2d-b531-08f22697c479" containerName="registry" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.167065 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.171051 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx"] Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.171353 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.171588 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.266769 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5464d49b-477d-4a25-95b2-77b1b83eb72b-config-volume\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.266849 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5464d49b-477d-4a25-95b2-77b1b83eb72b-secret-volume\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.266968 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk8t7\" (UniqueName: \"kubernetes.io/projected/5464d49b-477d-4a25-95b2-77b1b83eb72b-kube-api-access-sk8t7\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.368232 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5464d49b-477d-4a25-95b2-77b1b83eb72b-config-volume\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.368308 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5464d49b-477d-4a25-95b2-77b1b83eb72b-secret-volume\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.368355 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk8t7\" (UniqueName: \"kubernetes.io/projected/5464d49b-477d-4a25-95b2-77b1b83eb72b-kube-api-access-sk8t7\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.370479 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5464d49b-477d-4a25-95b2-77b1b83eb72b-config-volume\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.374579 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5464d49b-477d-4a25-95b2-77b1b83eb72b-secret-volume\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.395084 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk8t7\" (UniqueName: \"kubernetes.io/projected/5464d49b-477d-4a25-95b2-77b1b83eb72b-kube-api-access-sk8t7\") pod \"collect-profiles-29415615-vmggx\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.490615 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.654717 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx"] Dec 05 12:15:00 crc kubenswrapper[4807]: I1205 12:15:00.794022 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" event={"ID":"5464d49b-477d-4a25-95b2-77b1b83eb72b","Type":"ContainerStarted","Data":"73171d945eb2d08c140bef4b8f624aa339c2a1e172c865039c95734294ab16ab"} Dec 05 12:15:01 crc kubenswrapper[4807]: I1205 12:15:01.800922 4807 generic.go:334] "Generic (PLEG): container finished" podID="5464d49b-477d-4a25-95b2-77b1b83eb72b" containerID="57b6f4052ee6e888493ab88bdd0f6f961637e1cf5b0f3741b5cec71338c25c80" exitCode=0 Dec 05 12:15:01 crc kubenswrapper[4807]: I1205 12:15:01.800974 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" event={"ID":"5464d49b-477d-4a25-95b2-77b1b83eb72b","Type":"ContainerDied","Data":"57b6f4052ee6e888493ab88bdd0f6f961637e1cf5b0f3741b5cec71338c25c80"} Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.009736 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.206223 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk8t7\" (UniqueName: \"kubernetes.io/projected/5464d49b-477d-4a25-95b2-77b1b83eb72b-kube-api-access-sk8t7\") pod \"5464d49b-477d-4a25-95b2-77b1b83eb72b\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.206325 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5464d49b-477d-4a25-95b2-77b1b83eb72b-config-volume\") pod \"5464d49b-477d-4a25-95b2-77b1b83eb72b\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.206358 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5464d49b-477d-4a25-95b2-77b1b83eb72b-secret-volume\") pod \"5464d49b-477d-4a25-95b2-77b1b83eb72b\" (UID: \"5464d49b-477d-4a25-95b2-77b1b83eb72b\") " Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.207415 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5464d49b-477d-4a25-95b2-77b1b83eb72b-config-volume" (OuterVolumeSpecName: "config-volume") pod "5464d49b-477d-4a25-95b2-77b1b83eb72b" (UID: "5464d49b-477d-4a25-95b2-77b1b83eb72b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.211293 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5464d49b-477d-4a25-95b2-77b1b83eb72b-kube-api-access-sk8t7" (OuterVolumeSpecName: "kube-api-access-sk8t7") pod "5464d49b-477d-4a25-95b2-77b1b83eb72b" (UID: "5464d49b-477d-4a25-95b2-77b1b83eb72b"). InnerVolumeSpecName "kube-api-access-sk8t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.211298 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5464d49b-477d-4a25-95b2-77b1b83eb72b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5464d49b-477d-4a25-95b2-77b1b83eb72b" (UID: "5464d49b-477d-4a25-95b2-77b1b83eb72b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.307897 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk8t7\" (UniqueName: \"kubernetes.io/projected/5464d49b-477d-4a25-95b2-77b1b83eb72b-kube-api-access-sk8t7\") on node \"crc\" DevicePath \"\"" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.307947 4807 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5464d49b-477d-4a25-95b2-77b1b83eb72b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.307959 4807 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5464d49b-477d-4a25-95b2-77b1b83eb72b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.812185 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" event={"ID":"5464d49b-477d-4a25-95b2-77b1b83eb72b","Type":"ContainerDied","Data":"73171d945eb2d08c140bef4b8f624aa339c2a1e172c865039c95734294ab16ab"} Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.812228 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73171d945eb2d08c140bef4b8f624aa339c2a1e172c865039c95734294ab16ab" Dec 05 12:15:03 crc kubenswrapper[4807]: I1205 12:15:03.812266 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-vmggx" Dec 05 12:15:22 crc kubenswrapper[4807]: I1205 12:15:22.466214 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:15:22 crc kubenswrapper[4807]: I1205 12:15:22.466868 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:15:52 crc kubenswrapper[4807]: I1205 12:15:52.466449 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:15:52 crc kubenswrapper[4807]: I1205 12:15:52.467168 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:16:22 crc kubenswrapper[4807]: I1205 12:16:22.466645 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:16:22 crc kubenswrapper[4807]: I1205 12:16:22.467227 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:16:22 crc kubenswrapper[4807]: I1205 12:16:22.467279 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:16:22 crc kubenswrapper[4807]: I1205 12:16:22.467907 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4a9bf0256da38a66ac347b1c3d3069dc78bcbfdcd0d597ffe57a7702a56bd70"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:16:22 crc kubenswrapper[4807]: I1205 12:16:22.467970 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://d4a9bf0256da38a66ac347b1c3d3069dc78bcbfdcd0d597ffe57a7702a56bd70" gracePeriod=600 Dec 05 12:16:23 crc kubenswrapper[4807]: I1205 12:16:23.239673 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="d4a9bf0256da38a66ac347b1c3d3069dc78bcbfdcd0d597ffe57a7702a56bd70" exitCode=0 Dec 05 12:16:23 crc kubenswrapper[4807]: I1205 12:16:23.241005 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"d4a9bf0256da38a66ac347b1c3d3069dc78bcbfdcd0d597ffe57a7702a56bd70"} Dec 05 12:16:23 crc kubenswrapper[4807]: I1205 12:16:23.241052 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"e2671acad109a1034ec97e5cc79cbdeae0c327f0a6312fb0cd334defd0fc9ebe"} Dec 05 12:16:23 crc kubenswrapper[4807]: I1205 12:16:23.241074 4807 scope.go:117] "RemoveContainer" containerID="61fbdac8b8e46032c2f8db25d5c108f1020da47f0c78713ac4dad89d5fdc1819" Dec 05 12:18:22 crc kubenswrapper[4807]: I1205 12:18:22.466114 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:18:22 crc kubenswrapper[4807]: I1205 12:18:22.466644 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.305331 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l"] Dec 05 12:18:37 crc kubenswrapper[4807]: E1205 12:18:37.306156 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5464d49b-477d-4a25-95b2-77b1b83eb72b" containerName="collect-profiles" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.306175 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5464d49b-477d-4a25-95b2-77b1b83eb72b" containerName="collect-profiles" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.306302 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5464d49b-477d-4a25-95b2-77b1b83eb72b" containerName="collect-profiles" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.307238 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.310130 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.315517 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l"] Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.402421 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.402489 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.402580 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqp2m\" (UniqueName: \"kubernetes.io/projected/33a3d26a-4d53-4044-bbd6-bf6b704243ee-kube-api-access-fqp2m\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.503350 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.503441 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.503476 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqp2m\" (UniqueName: \"kubernetes.io/projected/33a3d26a-4d53-4044-bbd6-bf6b704243ee-kube-api-access-fqp2m\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.504188 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.504198 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.523181 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqp2m\" (UniqueName: \"kubernetes.io/projected/33a3d26a-4d53-4044-bbd6-bf6b704243ee-kube-api-access-fqp2m\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.629213 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:37 crc kubenswrapper[4807]: I1205 12:18:37.814423 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l"] Dec 05 12:18:38 crc kubenswrapper[4807]: I1205 12:18:38.087164 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" event={"ID":"33a3d26a-4d53-4044-bbd6-bf6b704243ee","Type":"ContainerStarted","Data":"5d26dfe44ed99fa80947328e03b9bd8e4d4872619082d9eac053863f11ec4b61"} Dec 05 12:18:38 crc kubenswrapper[4807]: I1205 12:18:38.087542 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" event={"ID":"33a3d26a-4d53-4044-bbd6-bf6b704243ee","Type":"ContainerStarted","Data":"70c7809801b2dddef6c1c843224c154a05880d7948a01f5f53e1aad44e11c8d5"} Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.093403 4807 generic.go:334] "Generic (PLEG): container finished" podID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerID="5d26dfe44ed99fa80947328e03b9bd8e4d4872619082d9eac053863f11ec4b61" exitCode=0 Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.093441 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" event={"ID":"33a3d26a-4d53-4044-bbd6-bf6b704243ee","Type":"ContainerDied","Data":"5d26dfe44ed99fa80947328e03b9bd8e4d4872619082d9eac053863f11ec4b61"} Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.097169 4807 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.411172 4807 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.672494 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qkxp9"] Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.673674 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.682143 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qkxp9"] Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.834168 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-catalog-content\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.834251 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-utilities\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.834339 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmvvd\" (UniqueName: \"kubernetes.io/projected/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-kube-api-access-hmvvd\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.935792 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-utilities\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.936351 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmvvd\" (UniqueName: \"kubernetes.io/projected/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-kube-api-access-hmvvd\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.936257 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-utilities\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.936450 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-catalog-content\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.936714 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-catalog-content\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.955142 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmvvd\" (UniqueName: \"kubernetes.io/projected/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-kube-api-access-hmvvd\") pod \"redhat-operators-qkxp9\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:39 crc kubenswrapper[4807]: I1205 12:18:39.990285 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:40 crc kubenswrapper[4807]: I1205 12:18:40.214006 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qkxp9"] Dec 05 12:18:40 crc kubenswrapper[4807]: W1205 12:18:40.323404 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f58a1b2_61ab_4f9a_8cdb_9a3ff0be212a.slice/crio-cfd384cea563856ee309bac59d549e557cf455b5ad59ac3cd12ee2683b97e483 WatchSource:0}: Error finding container cfd384cea563856ee309bac59d549e557cf455b5ad59ac3cd12ee2683b97e483: Status 404 returned error can't find the container with id cfd384cea563856ee309bac59d549e557cf455b5ad59ac3cd12ee2683b97e483 Dec 05 12:18:41 crc kubenswrapper[4807]: I1205 12:18:41.105481 4807 generic.go:334] "Generic (PLEG): container finished" podID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerID="f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4" exitCode=0 Dec 05 12:18:41 crc kubenswrapper[4807]: I1205 12:18:41.105576 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkxp9" event={"ID":"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a","Type":"ContainerDied","Data":"f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4"} Dec 05 12:18:41 crc kubenswrapper[4807]: I1205 12:18:41.105877 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkxp9" event={"ID":"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a","Type":"ContainerStarted","Data":"cfd384cea563856ee309bac59d549e557cf455b5ad59ac3cd12ee2683b97e483"} Dec 05 12:18:41 crc kubenswrapper[4807]: I1205 12:18:41.108190 4807 generic.go:334] "Generic (PLEG): container finished" podID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerID="63f6d66de6dd84f01f0cf88faa2f0d6862477d00d45f3eae2bd7357d22bcf2d1" exitCode=0 Dec 05 12:18:41 crc kubenswrapper[4807]: I1205 12:18:41.108229 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" event={"ID":"33a3d26a-4d53-4044-bbd6-bf6b704243ee","Type":"ContainerDied","Data":"63f6d66de6dd84f01f0cf88faa2f0d6862477d00d45f3eae2bd7357d22bcf2d1"} Dec 05 12:18:42 crc kubenswrapper[4807]: I1205 12:18:42.121655 4807 generic.go:334] "Generic (PLEG): container finished" podID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerID="269433bbf56924c3bdabcc388cebfa7f5dfcd4653290b88842e867cbc56a76a7" exitCode=0 Dec 05 12:18:42 crc kubenswrapper[4807]: I1205 12:18:42.121714 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" event={"ID":"33a3d26a-4d53-4044-bbd6-bf6b704243ee","Type":"ContainerDied","Data":"269433bbf56924c3bdabcc388cebfa7f5dfcd4653290b88842e867cbc56a76a7"} Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.128179 4807 generic.go:334] "Generic (PLEG): container finished" podID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerID="7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf" exitCode=0 Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.128285 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkxp9" event={"ID":"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a","Type":"ContainerDied","Data":"7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf"} Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.349500 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.481430 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqp2m\" (UniqueName: \"kubernetes.io/projected/33a3d26a-4d53-4044-bbd6-bf6b704243ee-kube-api-access-fqp2m\") pod \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.481495 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-util\") pod \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.481594 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-bundle\") pod \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\" (UID: \"33a3d26a-4d53-4044-bbd6-bf6b704243ee\") " Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.483956 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-bundle" (OuterVolumeSpecName: "bundle") pod "33a3d26a-4d53-4044-bbd6-bf6b704243ee" (UID: "33a3d26a-4d53-4044-bbd6-bf6b704243ee"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.487491 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a3d26a-4d53-4044-bbd6-bf6b704243ee-kube-api-access-fqp2m" (OuterVolumeSpecName: "kube-api-access-fqp2m") pod "33a3d26a-4d53-4044-bbd6-bf6b704243ee" (UID: "33a3d26a-4d53-4044-bbd6-bf6b704243ee"). InnerVolumeSpecName "kube-api-access-fqp2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.582385 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqp2m\" (UniqueName: \"kubernetes.io/projected/33a3d26a-4d53-4044-bbd6-bf6b704243ee-kube-api-access-fqp2m\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.582416 4807 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.821583 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-util" (OuterVolumeSpecName: "util") pod "33a3d26a-4d53-4044-bbd6-bf6b704243ee" (UID: "33a3d26a-4d53-4044-bbd6-bf6b704243ee"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:18:43 crc kubenswrapper[4807]: I1205 12:18:43.887017 4807 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33a3d26a-4d53-4044-bbd6-bf6b704243ee-util\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:44 crc kubenswrapper[4807]: I1205 12:18:44.138060 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" event={"ID":"33a3d26a-4d53-4044-bbd6-bf6b704243ee","Type":"ContainerDied","Data":"70c7809801b2dddef6c1c843224c154a05880d7948a01f5f53e1aad44e11c8d5"} Dec 05 12:18:44 crc kubenswrapper[4807]: I1205 12:18:44.138131 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70c7809801b2dddef6c1c843224c154a05880d7948a01f5f53e1aad44e11c8d5" Dec 05 12:18:44 crc kubenswrapper[4807]: I1205 12:18:44.138224 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l" Dec 05 12:18:45 crc kubenswrapper[4807]: I1205 12:18:45.145674 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkxp9" event={"ID":"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a","Type":"ContainerStarted","Data":"69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96"} Dec 05 12:18:45 crc kubenswrapper[4807]: I1205 12:18:45.170112 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qkxp9" podStartSLOduration=3.244742682 podStartE2EDuration="6.170093508s" podCreationTimestamp="2025-12-05 12:18:39 +0000 UTC" firstStartedPulling="2025-12-05 12:18:41.106597468 +0000 UTC m=+750.600460737" lastFinishedPulling="2025-12-05 12:18:44.031948294 +0000 UTC m=+753.525811563" observedRunningTime="2025-12-05 12:18:45.168055806 +0000 UTC m=+754.661919085" watchObservedRunningTime="2025-12-05 12:18:45.170093508 +0000 UTC m=+754.663956777" Dec 05 12:18:48 crc kubenswrapper[4807]: I1205 12:18:48.267404 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-84bd8"] Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.164199 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-controller" containerID="cri-o://7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.164256 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="northd" containerID="cri-o://529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.164316 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-node" containerID="cri-o://c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.164277 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="sbdb" containerID="cri-o://d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.164281 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="nbdb" containerID="cri-o://123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.164332 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.164339 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-acl-logging" containerID="cri-o://ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.211804 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" containerID="cri-o://bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac" gracePeriod=30 Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.990840 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:49 crc kubenswrapper[4807]: I1205 12:18:49.990911 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.170575 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/2.log" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.171286 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/1.log" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.171335 4807 generic.go:334] "Generic (PLEG): container finished" podID="379e7e5d-522d-409b-84db-fafbd1bff182" containerID="69b6cc6856985e969167fcd41b27cdb5f70f73283793841888c7c1db6caf3497" exitCode=2 Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.171435 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerDied","Data":"69b6cc6856985e969167fcd41b27cdb5f70f73283793841888c7c1db6caf3497"} Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.171504 4807 scope.go:117] "RemoveContainer" containerID="ef720901c82800990d384320627b7094fa6f8c00f3b334de9d0eab669cfa495f" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.172560 4807 scope.go:117] "RemoveContainer" containerID="69b6cc6856985e969167fcd41b27cdb5f70f73283793841888c7c1db6caf3497" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.173959 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovnkube-controller/3.log" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.176214 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovn-acl-logging/0.log" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.176727 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovn-controller/0.log" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177107 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac" exitCode=0 Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177183 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73" exitCode=0 Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177185 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac"} Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177241 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73"} Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177256 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9"} Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177197 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9" exitCode=0 Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177283 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276" exitCode=143 Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177299 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316" exitCode=143 Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177316 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276"} Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.177328 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316"} Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.269095 4807 scope.go:117] "RemoveContainer" containerID="3a1f5b7a94d932fafdc849809d82256568c03f4bece3f28455831a801fb6fdfe" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.652695 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovn-acl-logging/0.log" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.653569 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovn-controller/0.log" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.654358 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.777684 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-env-overrides\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.777825 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-slash\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.777907 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-ovn\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.777982 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-slash" (OuterVolumeSpecName: "host-slash") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778017 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778007 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-systemd-units\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778126 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-systemd\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778153 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-log-socket\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778175 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-netns\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778197 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-node-log\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778219 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-var-lib-openvswitch\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778243 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-config\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778257 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778284 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-script-lib\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778277 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778317 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778299 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-netd\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778371 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovn-node-metrics-cert\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778416 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-kubelet\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778461 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g748d\" (UniqueName: \"kubernetes.io/projected/a9bb753e-da01-4d90-95b2-5614ab510ec8-kube-api-access-g748d\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778489 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-ovn-kubernetes\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778512 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-etc-openvswitch\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778560 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-openvswitch\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778610 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-bin\") pod \"a9bb753e-da01-4d90-95b2-5614ab510ec8\" (UID: \"a9bb753e-da01-4d90-95b2-5614ab510ec8\") " Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778697 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-log-socket" (OuterVolumeSpecName: "log-socket") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778737 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778764 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-node-log" (OuterVolumeSpecName: "node-log") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778787 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778917 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779002 4807 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779011 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779032 4807 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-slash\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779061 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779087 4807 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779095 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779105 4807 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-log-socket\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779120 4807 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779113 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779130 4807 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-node-log\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779161 4807 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779173 4807 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.778961 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779221 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779450 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.779563 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.787784 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9bb753e-da01-4d90-95b2-5614ab510ec8-kube-api-access-g748d" (OuterVolumeSpecName: "kube-api-access-g748d") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "kube-api-access-g748d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.788066 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.806575 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a9bb753e-da01-4d90-95b2-5614ab510ec8" (UID: "a9bb753e-da01-4d90-95b2-5614ab510ec8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880078 4807 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880110 4807 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880120 4807 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880129 4807 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880143 4807 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880153 4807 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a9bb753e-da01-4d90-95b2-5614ab510ec8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880162 4807 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880170 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g748d\" (UniqueName: \"kubernetes.io/projected/a9bb753e-da01-4d90-95b2-5614ab510ec8-kube-api-access-g748d\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880180 4807 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880188 4807 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880196 4807 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.880204 4807 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a9bb753e-da01-4d90-95b2-5614ab510ec8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920032 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b5czf"] Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920224 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920236 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920248 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kubecfg-setup" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920254 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kubecfg-setup" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920261 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerName="util" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920267 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerName="util" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920273 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerName="pull" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920279 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerName="pull" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920289 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="nbdb" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920295 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="nbdb" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920306 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-acl-logging" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920311 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-acl-logging" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920320 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920326 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920335 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920340 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920347 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-node" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920352 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-node" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920360 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="sbdb" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920367 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="sbdb" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920374 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920380 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920387 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920393 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920399 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920404 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920411 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerName="extract" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920416 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerName="extract" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920425 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="northd" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920435 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="northd" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920565 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920574 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920583 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="northd" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920591 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a3d26a-4d53-4044-bbd6-bf6b704243ee" containerName="extract" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920598 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920606 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920614 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920622 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="kube-rbac-proxy-node" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920630 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="nbdb" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920638 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="sbdb" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920645 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920654 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovn-acl-logging" Dec 05 12:18:50 crc kubenswrapper[4807]: E1205 12:18:50.920750 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920757 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.920836 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerName="ovnkube-controller" Dec 05 12:18:50 crc kubenswrapper[4807]: I1205 12:18:50.922213 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083735 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-run-netns\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083777 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-etc-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083800 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-ovnkube-script-lib\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083828 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-run-ovn-kubernetes\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083844 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-node-log\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083863 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-ovn\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083879 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083897 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-cni-netd\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083918 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-log-socket\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083932 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-systemd-units\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083946 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083959 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-ovnkube-config\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083976 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6837171-50e7-41b7-860f-7ae13026d63c-ovn-node-metrics-cert\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.083992 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rfjk\" (UniqueName: \"kubernetes.io/projected/c6837171-50e7-41b7-860f-7ae13026d63c-kube-api-access-5rfjk\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.084007 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-systemd\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.084024 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-kubelet\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.084043 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-cni-bin\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.084074 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-slash\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.084095 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-env-overrides\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.084114 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-var-lib-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.087949 4807 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qkxp9" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="registry-server" probeResult="failure" output=< Dec 05 12:18:51 crc kubenswrapper[4807]: timeout: failed to connect service ":50051" within 1s Dec 05 12:18:51 crc kubenswrapper[4807]: > Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184797 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-run-netns\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184835 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-etc-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184858 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-ovnkube-script-lib\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184886 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-run-ovn-kubernetes\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184906 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-node-log\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184916 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-run-netns\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184929 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-ovn\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184961 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-ovn\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.184986 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185022 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-etc-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185032 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-cni-netd\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185062 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-log-socket\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185080 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-systemd-units\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185101 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185121 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-ovnkube-config\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185141 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6837171-50e7-41b7-860f-7ae13026d63c-ovn-node-metrics-cert\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185164 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rfjk\" (UniqueName: \"kubernetes.io/projected/c6837171-50e7-41b7-860f-7ae13026d63c-kube-api-access-5rfjk\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185183 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-systemd\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185199 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-kubelet\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185219 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-cni-bin\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185265 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-slash\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185283 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-env-overrides\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185299 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-var-lib-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185376 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-var-lib-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185401 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-openvswitch\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185422 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-cni-netd\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185441 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-log-socket\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185460 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-systemd-units\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185478 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185702 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-ovnkube-script-lib\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185737 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-run-ovn-kubernetes\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185759 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-node-log\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185780 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-kubelet\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.185998 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-ovnkube-config\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.186031 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-cni-bin\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.186052 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-host-slash\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.186209 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6837171-50e7-41b7-860f-7ae13026d63c-run-systemd\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.187041 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6837171-50e7-41b7-860f-7ae13026d63c-env-overrides\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.190964 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovn-acl-logging/0.log" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.191108 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6837171-50e7-41b7-860f-7ae13026d63c-ovn-node-metrics-cert\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.194803 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-84bd8_a9bb753e-da01-4d90-95b2-5614ab510ec8/ovn-controller/0.log" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195280 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5" exitCode=0 Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195309 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4" exitCode=0 Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195342 4807 generic.go:334] "Generic (PLEG): container finished" podID="a9bb753e-da01-4d90-95b2-5614ab510ec8" containerID="529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a" exitCode=0 Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195390 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5"} Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195411 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4"} Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195421 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a"} Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195430 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" event={"ID":"a9bb753e-da01-4d90-95b2-5614ab510ec8","Type":"ContainerDied","Data":"72e81ff6b8914de874641ae53375e42e97c220d16b2d2d1bef37443420f0d19e"} Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195448 4807 scope.go:117] "RemoveContainer" containerID="bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.195569 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-84bd8" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.202238 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sq7cs_379e7e5d-522d-409b-84db-fafbd1bff182/kube-multus/2.log" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.202301 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sq7cs" event={"ID":"379e7e5d-522d-409b-84db-fafbd1bff182","Type":"ContainerStarted","Data":"a942853e3eddf2c3574b226788046173c6870062e692f3b6f3f2f5b96d416246"} Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.213756 4807 scope.go:117] "RemoveContainer" containerID="d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.233438 4807 scope.go:117] "RemoveContainer" containerID="123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.249799 4807 scope.go:117] "RemoveContainer" containerID="529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.256714 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rfjk\" (UniqueName: \"kubernetes.io/projected/c6837171-50e7-41b7-860f-7ae13026d63c-kube-api-access-5rfjk\") pod \"ovnkube-node-b5czf\" (UID: \"c6837171-50e7-41b7-860f-7ae13026d63c\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.277923 4807 scope.go:117] "RemoveContainer" containerID="7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.315893 4807 scope.go:117] "RemoveContainer" containerID="c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.335853 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-84bd8"] Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.350312 4807 scope.go:117] "RemoveContainer" containerID="ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.377813 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-84bd8"] Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.378687 4807 scope.go:117] "RemoveContainer" containerID="7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.405702 4807 scope.go:117] "RemoveContainer" containerID="784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.430723 4807 scope.go:117] "RemoveContainer" containerID="bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.431240 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac\": container with ID starting with bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac not found: ID does not exist" containerID="bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.431278 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac"} err="failed to get container status \"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac\": rpc error: code = NotFound desc = could not find container \"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac\": container with ID starting with bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.431304 4807 scope.go:117] "RemoveContainer" containerID="d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.431902 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\": container with ID starting with d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5 not found: ID does not exist" containerID="d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.431941 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5"} err="failed to get container status \"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\": rpc error: code = NotFound desc = could not find container \"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\": container with ID starting with d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.431967 4807 scope.go:117] "RemoveContainer" containerID="123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.432259 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\": container with ID starting with 123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4 not found: ID does not exist" containerID="123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.432276 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4"} err="failed to get container status \"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\": rpc error: code = NotFound desc = could not find container \"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\": container with ID starting with 123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.432299 4807 scope.go:117] "RemoveContainer" containerID="529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.432511 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\": container with ID starting with 529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a not found: ID does not exist" containerID="529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.432550 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a"} err="failed to get container status \"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\": rpc error: code = NotFound desc = could not find container \"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\": container with ID starting with 529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.432566 4807 scope.go:117] "RemoveContainer" containerID="7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.432793 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\": container with ID starting with 7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73 not found: ID does not exist" containerID="7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.432810 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73"} err="failed to get container status \"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\": rpc error: code = NotFound desc = could not find container \"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\": container with ID starting with 7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.432822 4807 scope.go:117] "RemoveContainer" containerID="c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.432985 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\": container with ID starting with c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9 not found: ID does not exist" containerID="c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433001 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9"} err="failed to get container status \"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\": rpc error: code = NotFound desc = could not find container \"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\": container with ID starting with c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433011 4807 scope.go:117] "RemoveContainer" containerID="ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.433142 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\": container with ID starting with ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276 not found: ID does not exist" containerID="ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433158 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276"} err="failed to get container status \"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\": rpc error: code = NotFound desc = could not find container \"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\": container with ID starting with ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433168 4807 scope.go:117] "RemoveContainer" containerID="7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.433295 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\": container with ID starting with 7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316 not found: ID does not exist" containerID="7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433308 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316"} err="failed to get container status \"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\": rpc error: code = NotFound desc = could not find container \"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\": container with ID starting with 7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433318 4807 scope.go:117] "RemoveContainer" containerID="784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3" Dec 05 12:18:51 crc kubenswrapper[4807]: E1205 12:18:51.433470 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\": container with ID starting with 784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3 not found: ID does not exist" containerID="784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433484 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3"} err="failed to get container status \"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\": rpc error: code = NotFound desc = could not find container \"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\": container with ID starting with 784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433496 4807 scope.go:117] "RemoveContainer" containerID="bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433775 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac"} err="failed to get container status \"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac\": rpc error: code = NotFound desc = could not find container \"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac\": container with ID starting with bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.433789 4807 scope.go:117] "RemoveContainer" containerID="d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.434845 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5"} err="failed to get container status \"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\": rpc error: code = NotFound desc = could not find container \"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\": container with ID starting with d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.434859 4807 scope.go:117] "RemoveContainer" containerID="123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.435539 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4"} err="failed to get container status \"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\": rpc error: code = NotFound desc = could not find container \"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\": container with ID starting with 123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.435557 4807 scope.go:117] "RemoveContainer" containerID="529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.438345 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a"} err="failed to get container status \"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\": rpc error: code = NotFound desc = could not find container \"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\": container with ID starting with 529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.438374 4807 scope.go:117] "RemoveContainer" containerID="7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.441289 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73"} err="failed to get container status \"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\": rpc error: code = NotFound desc = could not find container \"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\": container with ID starting with 7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.441325 4807 scope.go:117] "RemoveContainer" containerID="c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.446649 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9"} err="failed to get container status \"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\": rpc error: code = NotFound desc = could not find container \"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\": container with ID starting with c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.446683 4807 scope.go:117] "RemoveContainer" containerID="ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.447072 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276"} err="failed to get container status \"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\": rpc error: code = NotFound desc = could not find container \"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\": container with ID starting with ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.447098 4807 scope.go:117] "RemoveContainer" containerID="7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.447649 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316"} err="failed to get container status \"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\": rpc error: code = NotFound desc = could not find container \"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\": container with ID starting with 7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.447666 4807 scope.go:117] "RemoveContainer" containerID="784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448053 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3"} err="failed to get container status \"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\": rpc error: code = NotFound desc = could not find container \"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\": container with ID starting with 784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448074 4807 scope.go:117] "RemoveContainer" containerID="bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448251 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac"} err="failed to get container status \"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac\": rpc error: code = NotFound desc = could not find container \"bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac\": container with ID starting with bbe0a698164ef316ac6152c63658cb287b79660230eee931dfe4ab246d8155ac not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448270 4807 scope.go:117] "RemoveContainer" containerID="d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448500 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5"} err="failed to get container status \"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\": rpc error: code = NotFound desc = could not find container \"d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5\": container with ID starting with d4fcff9c4c85bafcf1b4ab13a19dd077f6b998c3e0d055ff543237783169c1d5 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448516 4807 scope.go:117] "RemoveContainer" containerID="123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448920 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4"} err="failed to get container status \"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\": rpc error: code = NotFound desc = could not find container \"123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4\": container with ID starting with 123904a379b85a137246f61c530f5e0787f771a73cd44f09ff2701083044f5a4 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.448957 4807 scope.go:117] "RemoveContainer" containerID="529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.449222 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a"} err="failed to get container status \"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\": rpc error: code = NotFound desc = could not find container \"529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a\": container with ID starting with 529ee66cd7043c662be1329e099d4d6bfedf80d25b3eba46ecf134af7e11b76a not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.449244 4807 scope.go:117] "RemoveContainer" containerID="7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.449507 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73"} err="failed to get container status \"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\": rpc error: code = NotFound desc = could not find container \"7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73\": container with ID starting with 7462fda2016ab65f20e5efbd7754a088e1f0f64c3a049ac4c9c203c9e5307f73 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.449540 4807 scope.go:117] "RemoveContainer" containerID="c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.449749 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9"} err="failed to get container status \"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\": rpc error: code = NotFound desc = could not find container \"c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9\": container with ID starting with c5a7df7b0cabfd706d8b6cf161edaffc1fa1d8d6b2815c099a1d9795bde0e8c9 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.449770 4807 scope.go:117] "RemoveContainer" containerID="ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.449991 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276"} err="failed to get container status \"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\": rpc error: code = NotFound desc = could not find container \"ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276\": container with ID starting with ce0b1da6a5abab2794b92c5529dabe86164b38053d7bbe114222bc51f7f94276 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.450007 4807 scope.go:117] "RemoveContainer" containerID="7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.450170 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316"} err="failed to get container status \"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\": rpc error: code = NotFound desc = could not find container \"7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316\": container with ID starting with 7bba9f7e7a7f7114ae0c30480361d4f7629115cc79b94488b63d1ce746f95316 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.450186 4807 scope.go:117] "RemoveContainer" containerID="784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.450403 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3"} err="failed to get container status \"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\": rpc error: code = NotFound desc = could not find container \"784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3\": container with ID starting with 784676b237b5967937552921d22298ca756eeb0f552ca0e294a3c971af2564c3 not found: ID does not exist" Dec 05 12:18:51 crc kubenswrapper[4807]: I1205 12:18:51.534723 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:51 crc kubenswrapper[4807]: W1205 12:18:51.554899 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6837171_50e7_41b7_860f_7ae13026d63c.slice/crio-ad1398ea6349cf88b7783005127c989a8d249d3db41a0a1c47863fbde72fb26f WatchSource:0}: Error finding container ad1398ea6349cf88b7783005127c989a8d249d3db41a0a1c47863fbde72fb26f: Status 404 returned error can't find the container with id ad1398ea6349cf88b7783005127c989a8d249d3db41a0a1c47863fbde72fb26f Dec 05 12:18:52 crc kubenswrapper[4807]: I1205 12:18:52.214406 4807 generic.go:334] "Generic (PLEG): container finished" podID="c6837171-50e7-41b7-860f-7ae13026d63c" containerID="924c32cf681c1959bc3908c271867e1d88e51b50650fbfe0f5712b3ecb0b9ebd" exitCode=0 Dec 05 12:18:52 crc kubenswrapper[4807]: I1205 12:18:52.214457 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerDied","Data":"924c32cf681c1959bc3908c271867e1d88e51b50650fbfe0f5712b3ecb0b9ebd"} Dec 05 12:18:52 crc kubenswrapper[4807]: I1205 12:18:52.214483 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"ad1398ea6349cf88b7783005127c989a8d249d3db41a0a1c47863fbde72fb26f"} Dec 05 12:18:52 crc kubenswrapper[4807]: I1205 12:18:52.466068 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:18:52 crc kubenswrapper[4807]: I1205 12:18:52.466132 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:18:53 crc kubenswrapper[4807]: I1205 12:18:53.222460 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"dad7edbb69241d3d55c0e60db0ac7fc0cdf143bed5c2c97128bad503d6c5d06d"} Dec 05 12:18:53 crc kubenswrapper[4807]: I1205 12:18:53.222786 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"ab53bad21a0f30865ac8829bce0ea5bb8ecae2fff291cdfee98ede93bcf2f610"} Dec 05 12:18:53 crc kubenswrapper[4807]: I1205 12:18:53.222801 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"9fed55c7c02704b03baa3475ce43ea5a74a067b0f58fcd1d4ac63d346911dc85"} Dec 05 12:18:53 crc kubenswrapper[4807]: I1205 12:18:53.222811 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"cc246d338f76cd37cf2f2c783c84e434cca4b58a375806cf96776f68f40837c4"} Dec 05 12:18:53 crc kubenswrapper[4807]: I1205 12:18:53.222821 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"2828316cfe45aa8e9c19934bda7007b2c7297d09aac525bc8bacd71207c6c209"} Dec 05 12:18:53 crc kubenswrapper[4807]: I1205 12:18:53.222829 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"a7bb1142b1bdfc418bfa045ddcb31a914a89b40581c6987e96b682e689f0303c"} Dec 05 12:18:53 crc kubenswrapper[4807]: I1205 12:18:53.243560 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9bb753e-da01-4d90-95b2-5614ab510ec8" path="/var/lib/kubelet/pods/a9bb753e-da01-4d90-95b2-5614ab510ec8/volumes" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.264045 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk"] Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.265092 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.267644 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.267800 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.267818 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-s9786" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.376478 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w"] Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.377196 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.382124 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-rz9kr" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.382457 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.387905 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg"] Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.388593 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.438814 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftkhb\" (UniqueName: \"kubernetes.io/projected/c216b0c0-8388-42f5-96ec-618c61fa131b-kube-api-access-ftkhb\") pod \"obo-prometheus-operator-668cf9dfbb-7pnfk\" (UID: \"c216b0c0-8388-42f5-96ec-618c61fa131b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.540404 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73c61937-cac2-4ad4-abe6-4030f1fcdc0d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg\" (UID: \"73c61937-cac2-4ad4-abe6-4030f1fcdc0d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.540456 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aef16c20-caa8-4a1b-8b8a-df70e1588307-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w\" (UID: \"aef16c20-caa8-4a1b-8b8a-df70e1588307\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.540552 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73c61937-cac2-4ad4-abe6-4030f1fcdc0d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg\" (UID: \"73c61937-cac2-4ad4-abe6-4030f1fcdc0d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.540598 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aef16c20-caa8-4a1b-8b8a-df70e1588307-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w\" (UID: \"aef16c20-caa8-4a1b-8b8a-df70e1588307\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.540648 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftkhb\" (UniqueName: \"kubernetes.io/projected/c216b0c0-8388-42f5-96ec-618c61fa131b-kube-api-access-ftkhb\") pod \"obo-prometheus-operator-668cf9dfbb-7pnfk\" (UID: \"c216b0c0-8388-42f5-96ec-618c61fa131b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.565103 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftkhb\" (UniqueName: \"kubernetes.io/projected/c216b0c0-8388-42f5-96ec-618c61fa131b-kube-api-access-ftkhb\") pod \"obo-prometheus-operator-668cf9dfbb-7pnfk\" (UID: \"c216b0c0-8388-42f5-96ec-618c61fa131b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.591510 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.614178 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(15f66a0963d5c61484b1985c7ccafce3aa6d73c17a88888080a6a75f96739590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.614244 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(15f66a0963d5c61484b1985c7ccafce3aa6d73c17a88888080a6a75f96739590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.614266 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(15f66a0963d5c61484b1985c7ccafce3aa6d73c17a88888080a6a75f96739590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.614300 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators(c216b0c0-8388-42f5-96ec-618c61fa131b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators(c216b0c0-8388-42f5-96ec-618c61fa131b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(15f66a0963d5c61484b1985c7ccafce3aa6d73c17a88888080a6a75f96739590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" podUID="c216b0c0-8388-42f5-96ec-618c61fa131b" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.641938 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aef16c20-caa8-4a1b-8b8a-df70e1588307-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w\" (UID: \"aef16c20-caa8-4a1b-8b8a-df70e1588307\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.642020 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73c61937-cac2-4ad4-abe6-4030f1fcdc0d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg\" (UID: \"73c61937-cac2-4ad4-abe6-4030f1fcdc0d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.642065 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aef16c20-caa8-4a1b-8b8a-df70e1588307-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w\" (UID: \"aef16c20-caa8-4a1b-8b8a-df70e1588307\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.642121 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73c61937-cac2-4ad4-abe6-4030f1fcdc0d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg\" (UID: \"73c61937-cac2-4ad4-abe6-4030f1fcdc0d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.655094 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73c61937-cac2-4ad4-abe6-4030f1fcdc0d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg\" (UID: \"73c61937-cac2-4ad4-abe6-4030f1fcdc0d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.655141 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aef16c20-caa8-4a1b-8b8a-df70e1588307-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w\" (UID: \"aef16c20-caa8-4a1b-8b8a-df70e1588307\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.655152 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73c61937-cac2-4ad4-abe6-4030f1fcdc0d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg\" (UID: \"73c61937-cac2-4ad4-abe6-4030f1fcdc0d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.655201 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aef16c20-caa8-4a1b-8b8a-df70e1588307-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w\" (UID: \"aef16c20-caa8-4a1b-8b8a-df70e1588307\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.696003 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: I1205 12:18:55.707964 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.721737 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(8e7a85ca4473d43658cb3d5ef41bee8a30f1e7980d48901f0e677ee20a97b18b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.721822 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(8e7a85ca4473d43658cb3d5ef41bee8a30f1e7980d48901f0e677ee20a97b18b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.721844 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(8e7a85ca4473d43658cb3d5ef41bee8a30f1e7980d48901f0e677ee20a97b18b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.721914 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators(aef16c20-caa8-4a1b-8b8a-df70e1588307)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators(aef16c20-caa8-4a1b-8b8a-df70e1588307)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(8e7a85ca4473d43658cb3d5ef41bee8a30f1e7980d48901f0e677ee20a97b18b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" podUID="aef16c20-caa8-4a1b-8b8a-df70e1588307" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.734437 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(9ebdd61254218de919dc63b3def8b2ed8280e251325af510f340f535b8ef108b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.734518 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(9ebdd61254218de919dc63b3def8b2ed8280e251325af510f340f535b8ef108b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.734557 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(9ebdd61254218de919dc63b3def8b2ed8280e251325af510f340f535b8ef108b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:55 crc kubenswrapper[4807]: E1205 12:18:55.734622 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators(73c61937-cac2-4ad4-abe6-4030f1fcdc0d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators(73c61937-cac2-4ad4-abe6-4030f1fcdc0d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(9ebdd61254218de919dc63b3def8b2ed8280e251325af510f340f535b8ef108b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" podUID="73c61937-cac2-4ad4-abe6-4030f1fcdc0d" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.388266 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-fvbbz"] Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.389164 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.396826 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.398285 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-jn4zj" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.504504 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hd7g8"] Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.505164 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.520502 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-dv6k6" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.553912 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqg6x\" (UniqueName: \"kubernetes.io/projected/d571b11d-03a4-4879-9bf7-956322d80939-kube-api-access-bqg6x\") pod \"observability-operator-d8bb48f5d-fvbbz\" (UID: \"d571b11d-03a4-4879-9bf7-956322d80939\") " pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.553976 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d571b11d-03a4-4879-9bf7-956322d80939-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-fvbbz\" (UID: \"d571b11d-03a4-4879-9bf7-956322d80939\") " pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.654926 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqg6x\" (UniqueName: \"kubernetes.io/projected/d571b11d-03a4-4879-9bf7-956322d80939-kube-api-access-bqg6x\") pod \"observability-operator-d8bb48f5d-fvbbz\" (UID: \"d571b11d-03a4-4879-9bf7-956322d80939\") " pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.654987 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d571b11d-03a4-4879-9bf7-956322d80939-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-fvbbz\" (UID: \"d571b11d-03a4-4879-9bf7-956322d80939\") " pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.655044 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/78a728e0-7f03-44b3-8a15-bf94733000b2-openshift-service-ca\") pod \"perses-operator-5446b9c989-hd7g8\" (UID: \"78a728e0-7f03-44b3-8a15-bf94733000b2\") " pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.655087 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrr5s\" (UniqueName: \"kubernetes.io/projected/78a728e0-7f03-44b3-8a15-bf94733000b2-kube-api-access-qrr5s\") pod \"perses-operator-5446b9c989-hd7g8\" (UID: \"78a728e0-7f03-44b3-8a15-bf94733000b2\") " pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.660195 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d571b11d-03a4-4879-9bf7-956322d80939-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-fvbbz\" (UID: \"d571b11d-03a4-4879-9bf7-956322d80939\") " pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.683339 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqg6x\" (UniqueName: \"kubernetes.io/projected/d571b11d-03a4-4879-9bf7-956322d80939-kube-api-access-bqg6x\") pod \"observability-operator-d8bb48f5d-fvbbz\" (UID: \"d571b11d-03a4-4879-9bf7-956322d80939\") " pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.707204 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.744854 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(a4e182a104f6bea0e5b1b895219cfc41f8ae57856355077b7e5bc6480a6610a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.744948 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(a4e182a104f6bea0e5b1b895219cfc41f8ae57856355077b7e5bc6480a6610a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.744998 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(a4e182a104f6bea0e5b1b895219cfc41f8ae57856355077b7e5bc6480a6610a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.745077 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-fvbbz_openshift-operators(d571b11d-03a4-4879-9bf7-956322d80939)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-fvbbz_openshift-operators(d571b11d-03a4-4879-9bf7-956322d80939)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(a4e182a104f6bea0e5b1b895219cfc41f8ae57856355077b7e5bc6480a6610a6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" podUID="d571b11d-03a4-4879-9bf7-956322d80939" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.756182 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/78a728e0-7f03-44b3-8a15-bf94733000b2-openshift-service-ca\") pod \"perses-operator-5446b9c989-hd7g8\" (UID: \"78a728e0-7f03-44b3-8a15-bf94733000b2\") " pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.756268 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrr5s\" (UniqueName: \"kubernetes.io/projected/78a728e0-7f03-44b3-8a15-bf94733000b2-kube-api-access-qrr5s\") pod \"perses-operator-5446b9c989-hd7g8\" (UID: \"78a728e0-7f03-44b3-8a15-bf94733000b2\") " pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.757151 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/78a728e0-7f03-44b3-8a15-bf94733000b2-openshift-service-ca\") pod \"perses-operator-5446b9c989-hd7g8\" (UID: \"78a728e0-7f03-44b3-8a15-bf94733000b2\") " pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.786096 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrr5s\" (UniqueName: \"kubernetes.io/projected/78a728e0-7f03-44b3-8a15-bf94733000b2-kube-api-access-qrr5s\") pod \"perses-operator-5446b9c989-hd7g8\" (UID: \"78a728e0-7f03-44b3-8a15-bf94733000b2\") " pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: I1205 12:18:56.817732 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.837856 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(9706fd9a508988a3b136c125c0933af0ae68d4969d70c9fb367578c774a49491): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.837932 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(9706fd9a508988a3b136c125c0933af0ae68d4969d70c9fb367578c774a49491): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.837957 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(9706fd9a508988a3b136c125c0933af0ae68d4969d70c9fb367578c774a49491): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:56 crc kubenswrapper[4807]: E1205 12:18:56.838012 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-hd7g8_openshift-operators(78a728e0-7f03-44b3-8a15-bf94733000b2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-hd7g8_openshift-operators(78a728e0-7f03-44b3-8a15-bf94733000b2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(9706fd9a508988a3b136c125c0933af0ae68d4969d70c9fb367578c774a49491): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" podUID="78a728e0-7f03-44b3-8a15-bf94733000b2" Dec 05 12:18:57 crc kubenswrapper[4807]: I1205 12:18:57.243908 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"ea6c16dba86a4cb751d72e639fb8cc74d0a38e424798f2a00292603ec34bd739"} Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.266883 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" event={"ID":"c6837171-50e7-41b7-860f-7ae13026d63c","Type":"ContainerStarted","Data":"0e18fb058570591d44ccda1aa23c3dd438c4672858178321be7c442c5f77d5c3"} Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.269027 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.269059 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.269072 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.297039 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.302198 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.305021 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" podStartSLOduration=9.305011602 podStartE2EDuration="9.305011602s" podCreationTimestamp="2025-12-05 12:18:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:18:59.303283589 +0000 UTC m=+768.797146858" watchObservedRunningTime="2025-12-05 12:18:59.305011602 +0000 UTC m=+768.798874871" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.756716 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk"] Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.756942 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.757705 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.761848 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-fvbbz"] Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.762024 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.762684 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.781643 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg"] Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.781903 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.782437 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.785552 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hd7g8"] Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.785697 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.786138 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.788807 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w"] Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.788927 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.788963 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(f75a1cc3ec390b1e34323175eb11263dfbbd01aa41d144cd2aed240d61398738): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.789021 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(f75a1cc3ec390b1e34323175eb11263dfbbd01aa41d144cd2aed240d61398738): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.789051 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(f75a1cc3ec390b1e34323175eb11263dfbbd01aa41d144cd2aed240d61398738): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.789096 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators(c216b0c0-8388-42f5-96ec-618c61fa131b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators(c216b0c0-8388-42f5-96ec-618c61fa131b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-7pnfk_openshift-operators_c216b0c0-8388-42f5-96ec-618c61fa131b_0(f75a1cc3ec390b1e34323175eb11263dfbbd01aa41d144cd2aed240d61398738): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" podUID="c216b0c0-8388-42f5-96ec-618c61fa131b" Dec 05 12:18:59 crc kubenswrapper[4807]: I1205 12:18:59.789346 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.846776 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(4b416b24f5ea8874692e3b925175468aa0666745a2d52780005423a425289c0d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.846850 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(4b416b24f5ea8874692e3b925175468aa0666745a2d52780005423a425289c0d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.846878 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(4b416b24f5ea8874692e3b925175468aa0666745a2d52780005423a425289c0d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.846937 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-fvbbz_openshift-operators(d571b11d-03a4-4879-9bf7-956322d80939)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-fvbbz_openshift-operators(d571b11d-03a4-4879-9bf7-956322d80939)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fvbbz_openshift-operators_d571b11d-03a4-4879-9bf7-956322d80939_0(4b416b24f5ea8874692e3b925175468aa0666745a2d52780005423a425289c0d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" podUID="d571b11d-03a4-4879-9bf7-956322d80939" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.856583 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(35cfb1c5068f596a561f0f73d5506cd8a5a4c5d7644972eac0b39e43fad5c5b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.856663 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(35cfb1c5068f596a561f0f73d5506cd8a5a4c5d7644972eac0b39e43fad5c5b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.856689 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(35cfb1c5068f596a561f0f73d5506cd8a5a4c5d7644972eac0b39e43fad5c5b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.856751 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators(aef16c20-caa8-4a1b-8b8a-df70e1588307)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators(aef16c20-caa8-4a1b-8b8a-df70e1588307)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_openshift-operators_aef16c20-caa8-4a1b-8b8a-df70e1588307_0(35cfb1c5068f596a561f0f73d5506cd8a5a4c5d7644972eac0b39e43fad5c5b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" podUID="aef16c20-caa8-4a1b-8b8a-df70e1588307" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.863185 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(230c73f54cf50479410d34fd710b34484080446ab546164c33899f92af438544): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.863283 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(230c73f54cf50479410d34fd710b34484080446ab546164c33899f92af438544): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.863309 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(230c73f54cf50479410d34fd710b34484080446ab546164c33899f92af438544): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.863362 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-hd7g8_openshift-operators(78a728e0-7f03-44b3-8a15-bf94733000b2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-hd7g8_openshift-operators(78a728e0-7f03-44b3-8a15-bf94733000b2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-hd7g8_openshift-operators_78a728e0-7f03-44b3-8a15-bf94733000b2_0(230c73f54cf50479410d34fd710b34484080446ab546164c33899f92af438544): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" podUID="78a728e0-7f03-44b3-8a15-bf94733000b2" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.869864 4807 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(898179032ad9e4e8950f7e99e8d13302ae8679a5087a29670bad5e0751cee3ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.869915 4807 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(898179032ad9e4e8950f7e99e8d13302ae8679a5087a29670bad5e0751cee3ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.869942 4807 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(898179032ad9e4e8950f7e99e8d13302ae8679a5087a29670bad5e0751cee3ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:18:59 crc kubenswrapper[4807]: E1205 12:18:59.869992 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators(73c61937-cac2-4ad4-abe6-4030f1fcdc0d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators(73c61937-cac2-4ad4-abe6-4030f1fcdc0d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_openshift-operators_73c61937-cac2-4ad4-abe6-4030f1fcdc0d_0(898179032ad9e4e8950f7e99e8d13302ae8679a5087a29670bad5e0751cee3ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" podUID="73c61937-cac2-4ad4-abe6-4030f1fcdc0d" Dec 05 12:19:00 crc kubenswrapper[4807]: I1205 12:19:00.040061 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:19:00 crc kubenswrapper[4807]: I1205 12:19:00.098302 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:19:00 crc kubenswrapper[4807]: I1205 12:19:00.288814 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qkxp9"] Dec 05 12:19:01 crc kubenswrapper[4807]: I1205 12:19:01.276073 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qkxp9" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="registry-server" containerID="cri-o://69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96" gracePeriod=2 Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.250002 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.289639 4807 generic.go:334] "Generic (PLEG): container finished" podID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerID="69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96" exitCode=0 Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.289692 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkxp9" event={"ID":"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a","Type":"ContainerDied","Data":"69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96"} Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.289726 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkxp9" event={"ID":"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a","Type":"ContainerDied","Data":"cfd384cea563856ee309bac59d549e557cf455b5ad59ac3cd12ee2683b97e483"} Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.289746 4807 scope.go:117] "RemoveContainer" containerID="69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.289884 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkxp9" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.326548 4807 scope.go:117] "RemoveContainer" containerID="7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.346460 4807 scope.go:117] "RemoveContainer" containerID="f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.370665 4807 scope.go:117] "RemoveContainer" containerID="69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96" Dec 05 12:19:02 crc kubenswrapper[4807]: E1205 12:19:02.371128 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96\": container with ID starting with 69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96 not found: ID does not exist" containerID="69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.371203 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96"} err="failed to get container status \"69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96\": rpc error: code = NotFound desc = could not find container \"69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96\": container with ID starting with 69a71d1cf6fec739302e0a1c49753cba11989415e422c2231b6063a3059dba96 not found: ID does not exist" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.371254 4807 scope.go:117] "RemoveContainer" containerID="7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf" Dec 05 12:19:02 crc kubenswrapper[4807]: E1205 12:19:02.371581 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf\": container with ID starting with 7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf not found: ID does not exist" containerID="7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.371619 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf"} err="failed to get container status \"7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf\": rpc error: code = NotFound desc = could not find container \"7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf\": container with ID starting with 7eb7738a2ce4733e09e6a9cf4d693fb86050d7136b6256c7a027606caec652bf not found: ID does not exist" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.371645 4807 scope.go:117] "RemoveContainer" containerID="f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4" Dec 05 12:19:02 crc kubenswrapper[4807]: E1205 12:19:02.371871 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4\": container with ID starting with f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4 not found: ID does not exist" containerID="f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.371896 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4"} err="failed to get container status \"f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4\": rpc error: code = NotFound desc = could not find container \"f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4\": container with ID starting with f8eeb938a92c8bc686820a328dc2be8e4c65043e3d28ebeb2bcbe8b0dbd2d3b4 not found: ID does not exist" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.426918 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-utilities\") pod \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.426962 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmvvd\" (UniqueName: \"kubernetes.io/projected/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-kube-api-access-hmvvd\") pod \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.427076 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-catalog-content\") pod \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\" (UID: \"4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a\") " Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.427839 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-utilities" (OuterVolumeSpecName: "utilities") pod "4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" (UID: "4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.434983 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-kube-api-access-hmvvd" (OuterVolumeSpecName: "kube-api-access-hmvvd") pod "4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" (UID: "4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a"). InnerVolumeSpecName "kube-api-access-hmvvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.528769 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.528808 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmvvd\" (UniqueName: \"kubernetes.io/projected/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-kube-api-access-hmvvd\") on node \"crc\" DevicePath \"\"" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.565568 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" (UID: "4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.616371 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qkxp9"] Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.626284 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qkxp9"] Dec 05 12:19:02 crc kubenswrapper[4807]: I1205 12:19:02.630381 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:19:03 crc kubenswrapper[4807]: I1205 12:19:03.241629 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" path="/var/lib/kubelet/pods/4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a/volumes" Dec 05 12:19:12 crc kubenswrapper[4807]: I1205 12:19:12.235409 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:19:12 crc kubenswrapper[4807]: I1205 12:19:12.235595 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:19:12 crc kubenswrapper[4807]: I1205 12:19:12.236543 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" Dec 05 12:19:12 crc kubenswrapper[4807]: I1205 12:19:12.236691 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:19:12 crc kubenswrapper[4807]: I1205 12:19:12.487248 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hd7g8"] Dec 05 12:19:12 crc kubenswrapper[4807]: W1205 12:19:12.490984 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78a728e0_7f03_44b3_8a15_bf94733000b2.slice/crio-700b372f9924c85df38cfc0cac9f956d377985546a5289a09449462dcd50b1c1 WatchSource:0}: Error finding container 700b372f9924c85df38cfc0cac9f956d377985546a5289a09449462dcd50b1c1: Status 404 returned error can't find the container with id 700b372f9924c85df38cfc0cac9f956d377985546a5289a09449462dcd50b1c1 Dec 05 12:19:12 crc kubenswrapper[4807]: I1205 12:19:12.643196 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk"] Dec 05 12:19:12 crc kubenswrapper[4807]: W1205 12:19:12.645860 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc216b0c0_8388_42f5_96ec_618c61fa131b.slice/crio-ed227e7e184f58c2ec2d335f2b54a77237ce79f3058752fa91398f5e51c2a306 WatchSource:0}: Error finding container ed227e7e184f58c2ec2d335f2b54a77237ce79f3058752fa91398f5e51c2a306: Status 404 returned error can't find the container with id ed227e7e184f58c2ec2d335f2b54a77237ce79f3058752fa91398f5e51c2a306 Dec 05 12:19:13 crc kubenswrapper[4807]: I1205 12:19:13.235660 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:19:13 crc kubenswrapper[4807]: I1205 12:19:13.237414 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" Dec 05 12:19:13 crc kubenswrapper[4807]: I1205 12:19:13.357613 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" event={"ID":"78a728e0-7f03-44b3-8a15-bf94733000b2","Type":"ContainerStarted","Data":"700b372f9924c85df38cfc0cac9f956d377985546a5289a09449462dcd50b1c1"} Dec 05 12:19:13 crc kubenswrapper[4807]: I1205 12:19:13.360177 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" event={"ID":"c216b0c0-8388-42f5-96ec-618c61fa131b","Type":"ContainerStarted","Data":"ed227e7e184f58c2ec2d335f2b54a77237ce79f3058752fa91398f5e51c2a306"} Dec 05 12:19:13 crc kubenswrapper[4807]: I1205 12:19:13.479723 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w"] Dec 05 12:19:13 crc kubenswrapper[4807]: W1205 12:19:13.489671 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaef16c20_caa8_4a1b_8b8a_df70e1588307.slice/crio-4a4e62dcd36da6aafc9c424a167442f7f4dddb34e3355f4b5d4efc969d80c5ad WatchSource:0}: Error finding container 4a4e62dcd36da6aafc9c424a167442f7f4dddb34e3355f4b5d4efc969d80c5ad: Status 404 returned error can't find the container with id 4a4e62dcd36da6aafc9c424a167442f7f4dddb34e3355f4b5d4efc969d80c5ad Dec 05 12:19:14 crc kubenswrapper[4807]: I1205 12:19:14.234617 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:19:14 crc kubenswrapper[4807]: I1205 12:19:14.235420 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:19:14 crc kubenswrapper[4807]: I1205 12:19:14.372433 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" event={"ID":"aef16c20-caa8-4a1b-8b8a-df70e1588307","Type":"ContainerStarted","Data":"4a4e62dcd36da6aafc9c424a167442f7f4dddb34e3355f4b5d4efc969d80c5ad"} Dec 05 12:19:14 crc kubenswrapper[4807]: I1205 12:19:14.538345 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-fvbbz"] Dec 05 12:19:15 crc kubenswrapper[4807]: I1205 12:19:15.238006 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:19:15 crc kubenswrapper[4807]: I1205 12:19:15.238947 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" Dec 05 12:19:15 crc kubenswrapper[4807]: I1205 12:19:15.400652 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" event={"ID":"d571b11d-03a4-4879-9bf7-956322d80939","Type":"ContainerStarted","Data":"2f1bc2748c5382db8c377b2ef88dbcf720976e6ebf19151420ccbc02a6240d24"} Dec 05 12:19:15 crc kubenswrapper[4807]: I1205 12:19:15.722209 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg"] Dec 05 12:19:16 crc kubenswrapper[4807]: I1205 12:19:16.414170 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" event={"ID":"73c61937-cac2-4ad4-abe6-4030f1fcdc0d","Type":"ContainerStarted","Data":"e0e292ca65ee877b352bbb52de174faa0e210a3f47ef3fb7d4a256e5f0de4399"} Dec 05 12:19:21 crc kubenswrapper[4807]: I1205 12:19:21.560246 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b5czf" Dec 05 12:19:22 crc kubenswrapper[4807]: I1205 12:19:22.466609 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:19:22 crc kubenswrapper[4807]: I1205 12:19:22.466992 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:19:22 crc kubenswrapper[4807]: I1205 12:19:22.467038 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:19:22 crc kubenswrapper[4807]: I1205 12:19:22.467903 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e2671acad109a1034ec97e5cc79cbdeae0c327f0a6312fb0cd334defd0fc9ebe"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:19:22 crc kubenswrapper[4807]: I1205 12:19:22.467972 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://e2671acad109a1034ec97e5cc79cbdeae0c327f0a6312fb0cd334defd0fc9ebe" gracePeriod=600 Dec 05 12:19:24 crc kubenswrapper[4807]: I1205 12:19:24.515624 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="e2671acad109a1034ec97e5cc79cbdeae0c327f0a6312fb0cd334defd0fc9ebe" exitCode=0 Dec 05 12:19:24 crc kubenswrapper[4807]: I1205 12:19:24.515665 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"e2671acad109a1034ec97e5cc79cbdeae0c327f0a6312fb0cd334defd0fc9ebe"} Dec 05 12:19:24 crc kubenswrapper[4807]: I1205 12:19:24.515696 4807 scope.go:117] "RemoveContainer" containerID="d4a9bf0256da38a66ac347b1c3d3069dc78bcbfdcd0d597ffe57a7702a56bd70" Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.532926 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" event={"ID":"78a728e0-7f03-44b3-8a15-bf94733000b2","Type":"ContainerStarted","Data":"bf18d3663831cf77562d09ecd003d0f6f94e163494d36f98656fed5399fd9895"} Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.534601 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.535369 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"9d6e7e624977c9e3e4da4ace9e23b2a19ff1d1ccbc78d45f4be87d54cc1e3e37"} Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.539085 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" event={"ID":"73c61937-cac2-4ad4-abe6-4030f1fcdc0d","Type":"ContainerStarted","Data":"d671eab9926734eb35d0f294ce3ea5f25d6be125ce83f6a4f6402677ea901a68"} Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.541175 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" event={"ID":"d571b11d-03a4-4879-9bf7-956322d80939","Type":"ContainerStarted","Data":"2831c3580b57282a7c8d88b0a90d365845fc0516e490901bb2ded2e9a163fdea"} Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.541559 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.542717 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" event={"ID":"aef16c20-caa8-4a1b-8b8a-df70e1588307","Type":"ContainerStarted","Data":"aeaa08a8001d87ea5acdcf1079a03bb8983e0702ec3f8dead0e06085e8b3398e"} Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.543065 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.544509 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" event={"ID":"c216b0c0-8388-42f5-96ec-618c61fa131b","Type":"ContainerStarted","Data":"c051164c2829138a25667302e0d836aa8eafd7c2b381d1ba6cbf382d67594982"} Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.560043 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" podStartSLOduration=17.333601785 podStartE2EDuration="31.560027832s" podCreationTimestamp="2025-12-05 12:18:56 +0000 UTC" firstStartedPulling="2025-12-05 12:19:12.493211955 +0000 UTC m=+781.987075234" lastFinishedPulling="2025-12-05 12:19:26.719638012 +0000 UTC m=+796.213501281" observedRunningTime="2025-12-05 12:19:27.557117409 +0000 UTC m=+797.050980679" watchObservedRunningTime="2025-12-05 12:19:27.560027832 +0000 UTC m=+797.053891101" Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.581872 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-7pnfk" podStartSLOduration=18.582496019 podStartE2EDuration="32.581855597s" podCreationTimestamp="2025-12-05 12:18:55 +0000 UTC" firstStartedPulling="2025-12-05 12:19:12.647976579 +0000 UTC m=+782.141839858" lastFinishedPulling="2025-12-05 12:19:26.647336157 +0000 UTC m=+796.141199436" observedRunningTime="2025-12-05 12:19:27.580921194 +0000 UTC m=+797.074784513" watchObservedRunningTime="2025-12-05 12:19:27.581855597 +0000 UTC m=+797.075718866" Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.613442 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg" podStartSLOduration=21.619113038 podStartE2EDuration="32.613414755s" podCreationTimestamp="2025-12-05 12:18:55 +0000 UTC" firstStartedPulling="2025-12-05 12:19:15.750996085 +0000 UTC m=+785.244859354" lastFinishedPulling="2025-12-05 12:19:26.745297802 +0000 UTC m=+796.239161071" observedRunningTime="2025-12-05 12:19:27.607742023 +0000 UTC m=+797.101605312" watchObservedRunningTime="2025-12-05 12:19:27.613414755 +0000 UTC m=+797.107278044" Dec 05 12:19:27 crc kubenswrapper[4807]: I1205 12:19:27.642626 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-fvbbz" podStartSLOduration=19.47400854 podStartE2EDuration="31.642598823s" podCreationTimestamp="2025-12-05 12:18:56 +0000 UTC" firstStartedPulling="2025-12-05 12:19:14.551141351 +0000 UTC m=+784.045004620" lastFinishedPulling="2025-12-05 12:19:26.719731624 +0000 UTC m=+796.213594903" observedRunningTime="2025-12-05 12:19:27.636101362 +0000 UTC m=+797.129964651" watchObservedRunningTime="2025-12-05 12:19:27.642598823 +0000 UTC m=+797.136462142" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.555772 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w" podStartSLOduration=28.416384248 podStartE2EDuration="41.555752416s" podCreationTimestamp="2025-12-05 12:18:55 +0000 UTC" firstStartedPulling="2025-12-05 12:19:13.492231116 +0000 UTC m=+782.986094385" lastFinishedPulling="2025-12-05 12:19:26.631599264 +0000 UTC m=+796.125462553" observedRunningTime="2025-12-05 12:19:27.669367452 +0000 UTC m=+797.163230781" watchObservedRunningTime="2025-12-05 12:19:36.555752416 +0000 UTC m=+806.049615685" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.557573 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm"] Dec 05 12:19:36 crc kubenswrapper[4807]: E1205 12:19:36.557768 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="extract-utilities" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.557787 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="extract-utilities" Dec 05 12:19:36 crc kubenswrapper[4807]: E1205 12:19:36.557800 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="registry-server" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.557808 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="registry-server" Dec 05 12:19:36 crc kubenswrapper[4807]: E1205 12:19:36.557825 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="extract-content" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.557832 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="extract-content" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.557972 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f58a1b2-61ab-4f9a-8cdb-9a3ff0be212a" containerName="registry-server" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.558855 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.561074 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.571642 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm"] Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.748264 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.748754 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.748813 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/471da076-74b7-4d22-b5ac-67d760a8ee53-kube-api-access-fj8h7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.821304 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-hd7g8" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.849552 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.849619 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/471da076-74b7-4d22-b5ac-67d760a8ee53-kube-api-access-fj8h7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.850360 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.850416 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.851641 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.874025 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/471da076-74b7-4d22-b5ac-67d760a8ee53-kube-api-access-fj8h7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:36 crc kubenswrapper[4807]: I1205 12:19:36.897933 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:37 crc kubenswrapper[4807]: I1205 12:19:37.155230 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm"] Dec 05 12:19:37 crc kubenswrapper[4807]: I1205 12:19:37.603666 4807 generic.go:334] "Generic (PLEG): container finished" podID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerID="e89520107976d8eefec74e00dcd20e314e8132cd2f01f653dd15dfbff6216e54" exitCode=0 Dec 05 12:19:37 crc kubenswrapper[4807]: I1205 12:19:37.603720 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" event={"ID":"471da076-74b7-4d22-b5ac-67d760a8ee53","Type":"ContainerDied","Data":"e89520107976d8eefec74e00dcd20e314e8132cd2f01f653dd15dfbff6216e54"} Dec 05 12:19:37 crc kubenswrapper[4807]: I1205 12:19:37.603751 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" event={"ID":"471da076-74b7-4d22-b5ac-67d760a8ee53","Type":"ContainerStarted","Data":"8ccd9c449ee108fad772b1613e7d8cb1d5a4a712dd5a918b2b481f6d164f471d"} Dec 05 12:19:39 crc kubenswrapper[4807]: I1205 12:19:39.614883 4807 generic.go:334] "Generic (PLEG): container finished" podID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerID="ccc1638a80d062cc2cc02d6454e0cb099ca974065e4275bd8781f0ba165d7f22" exitCode=0 Dec 05 12:19:39 crc kubenswrapper[4807]: I1205 12:19:39.614965 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" event={"ID":"471da076-74b7-4d22-b5ac-67d760a8ee53","Type":"ContainerDied","Data":"ccc1638a80d062cc2cc02d6454e0cb099ca974065e4275bd8781f0ba165d7f22"} Dec 05 12:19:40 crc kubenswrapper[4807]: I1205 12:19:40.622595 4807 generic.go:334] "Generic (PLEG): container finished" podID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerID="4d9a2285d0fef6fea7430822fc617405d58a360e9545187f56d823d5f98f9378" exitCode=0 Dec 05 12:19:40 crc kubenswrapper[4807]: I1205 12:19:40.622750 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" event={"ID":"471da076-74b7-4d22-b5ac-67d760a8ee53","Type":"ContainerDied","Data":"4d9a2285d0fef6fea7430822fc617405d58a360e9545187f56d823d5f98f9378"} Dec 05 12:19:41 crc kubenswrapper[4807]: I1205 12:19:41.844372 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.008923 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/471da076-74b7-4d22-b5ac-67d760a8ee53-kube-api-access-fj8h7\") pod \"471da076-74b7-4d22-b5ac-67d760a8ee53\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.009025 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-bundle\") pod \"471da076-74b7-4d22-b5ac-67d760a8ee53\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.009053 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-util\") pod \"471da076-74b7-4d22-b5ac-67d760a8ee53\" (UID: \"471da076-74b7-4d22-b5ac-67d760a8ee53\") " Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.009675 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-bundle" (OuterVolumeSpecName: "bundle") pod "471da076-74b7-4d22-b5ac-67d760a8ee53" (UID: "471da076-74b7-4d22-b5ac-67d760a8ee53"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.016766 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/471da076-74b7-4d22-b5ac-67d760a8ee53-kube-api-access-fj8h7" (OuterVolumeSpecName: "kube-api-access-fj8h7") pod "471da076-74b7-4d22-b5ac-67d760a8ee53" (UID: "471da076-74b7-4d22-b5ac-67d760a8ee53"). InnerVolumeSpecName "kube-api-access-fj8h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.023120 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-util" (OuterVolumeSpecName: "util") pod "471da076-74b7-4d22-b5ac-67d760a8ee53" (UID: "471da076-74b7-4d22-b5ac-67d760a8ee53"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.109792 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/471da076-74b7-4d22-b5ac-67d760a8ee53-kube-api-access-fj8h7\") on node \"crc\" DevicePath \"\"" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.109825 4807 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.109838 4807 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/471da076-74b7-4d22-b5ac-67d760a8ee53-util\") on node \"crc\" DevicePath \"\"" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.636676 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" event={"ID":"471da076-74b7-4d22-b5ac-67d760a8ee53","Type":"ContainerDied","Data":"8ccd9c449ee108fad772b1613e7d8cb1d5a4a712dd5a918b2b481f6d164f471d"} Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.636732 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ccd9c449ee108fad772b1613e7d8cb1d5a4a712dd5a918b2b481f6d164f471d" Dec 05 12:19:42 crc kubenswrapper[4807]: I1205 12:19:42.636763 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.425228 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9"] Dec 05 12:19:48 crc kubenswrapper[4807]: E1205 12:19:48.426066 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerName="pull" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.426082 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerName="pull" Dec 05 12:19:48 crc kubenswrapper[4807]: E1205 12:19:48.426100 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerName="extract" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.426109 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerName="extract" Dec 05 12:19:48 crc kubenswrapper[4807]: E1205 12:19:48.426120 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerName="util" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.426128 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerName="util" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.426249 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="471da076-74b7-4d22-b5ac-67d760a8ee53" containerName="extract" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.426783 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.429160 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.429242 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-psr55" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.429518 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.436844 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9"] Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.486161 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt44c\" (UniqueName: \"kubernetes.io/projected/647cdbda-7377-42a5-80d2-d67f6babcc58-kube-api-access-mt44c\") pod \"nmstate-operator-5b5b58f5c8-kbls9\" (UID: \"647cdbda-7377-42a5-80d2-d67f6babcc58\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.586960 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt44c\" (UniqueName: \"kubernetes.io/projected/647cdbda-7377-42a5-80d2-d67f6babcc58-kube-api-access-mt44c\") pod \"nmstate-operator-5b5b58f5c8-kbls9\" (UID: \"647cdbda-7377-42a5-80d2-d67f6babcc58\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.605178 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt44c\" (UniqueName: \"kubernetes.io/projected/647cdbda-7377-42a5-80d2-d67f6babcc58-kube-api-access-mt44c\") pod \"nmstate-operator-5b5b58f5c8-kbls9\" (UID: \"647cdbda-7377-42a5-80d2-d67f6babcc58\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.745430 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" Dec 05 12:19:48 crc kubenswrapper[4807]: I1205 12:19:48.996035 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9"] Dec 05 12:19:49 crc kubenswrapper[4807]: I1205 12:19:49.683039 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" event={"ID":"647cdbda-7377-42a5-80d2-d67f6babcc58","Type":"ContainerStarted","Data":"689b2c4651feb3b1cf54d57a4f679e339189c13adb9dfe4d856bc17e0650667d"} Dec 05 12:19:51 crc kubenswrapper[4807]: I1205 12:19:51.697106 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" event={"ID":"647cdbda-7377-42a5-80d2-d67f6babcc58","Type":"ContainerStarted","Data":"9d093f338541e62e9e95152fcae2674ab65c2ade27ab013dfa6c3a733ed4fe99"} Dec 05 12:19:51 crc kubenswrapper[4807]: I1205 12:19:51.712555 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-kbls9" podStartSLOduration=1.7164125970000002 podStartE2EDuration="3.712513339s" podCreationTimestamp="2025-12-05 12:19:48 +0000 UTC" firstStartedPulling="2025-12-05 12:19:49.047497178 +0000 UTC m=+818.541360447" lastFinishedPulling="2025-12-05 12:19:51.04359792 +0000 UTC m=+820.537461189" observedRunningTime="2025-12-05 12:19:51.708707355 +0000 UTC m=+821.202570644" watchObservedRunningTime="2025-12-05 12:19:51.712513339 +0000 UTC m=+821.206376608" Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.856192 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p442k"] Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.857576 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.866334 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-s5f28" Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.867216 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p442k"] Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.882831 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr"] Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.883665 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.890787 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.918551 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-hp4qg"] Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.919513 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.926249 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr"] Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.989034 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks"] Dec 05 12:19:57 crc kubenswrapper[4807]: I1205 12:19:57.999502 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.003634 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks"] Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.008909 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jsfgt" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.008927 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.009056 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.025612 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8blb2\" (UniqueName: \"kubernetes.io/projected/222fa6f9-b50a-4ef2-82fd-7b9d37841def-kube-api-access-8blb2\") pod \"nmstate-metrics-7f946cbc9-p442k\" (UID: \"222fa6f9-b50a-4ef2-82fd-7b9d37841def\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.025741 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-dbus-socket\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.025785 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-ovs-socket\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.025811 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6d436024-c16c-4428-81db-a144ed2deecc-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-tcgqr\" (UID: \"6d436024-c16c-4428-81db-a144ed2deecc\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.025837 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-nmstate-lock\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.025861 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8glrq\" (UniqueName: \"kubernetes.io/projected/6d436024-c16c-4428-81db-a144ed2deecc-kube-api-access-8glrq\") pod \"nmstate-webhook-5f6d4c5ccb-tcgqr\" (UID: \"6d436024-c16c-4428-81db-a144ed2deecc\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.025887 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p9kw\" (UniqueName: \"kubernetes.io/projected/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-kube-api-access-6p9kw\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127191 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/79eac56e-1d44-4f6b-875a-8f944ee63f7a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127243 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-ovs-socket\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127269 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6d436024-c16c-4428-81db-a144ed2deecc-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-tcgqr\" (UID: \"6d436024-c16c-4428-81db-a144ed2deecc\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127289 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-nmstate-lock\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127311 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/79eac56e-1d44-4f6b-875a-8f944ee63f7a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127328 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8glrq\" (UniqueName: \"kubernetes.io/projected/6d436024-c16c-4428-81db-a144ed2deecc-kube-api-access-8glrq\") pod \"nmstate-webhook-5f6d4c5ccb-tcgqr\" (UID: \"6d436024-c16c-4428-81db-a144ed2deecc\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127404 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-nmstate-lock\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127408 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p9kw\" (UniqueName: \"kubernetes.io/projected/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-kube-api-access-6p9kw\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127600 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8blb2\" (UniqueName: \"kubernetes.io/projected/222fa6f9-b50a-4ef2-82fd-7b9d37841def-kube-api-access-8blb2\") pod \"nmstate-metrics-7f946cbc9-p442k\" (UID: \"222fa6f9-b50a-4ef2-82fd-7b9d37841def\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127641 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sffbj\" (UniqueName: \"kubernetes.io/projected/79eac56e-1d44-4f6b-875a-8f944ee63f7a-kube-api-access-sffbj\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127770 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-ovs-socket\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127778 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-dbus-socket\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.127970 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-dbus-socket\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.135318 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6d436024-c16c-4428-81db-a144ed2deecc-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-tcgqr\" (UID: \"6d436024-c16c-4428-81db-a144ed2deecc\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.152997 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p9kw\" (UniqueName: \"kubernetes.io/projected/1ef69c53-a12b-4ae8-ae6f-baa299acfb6a-kube-api-access-6p9kw\") pod \"nmstate-handler-hp4qg\" (UID: \"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a\") " pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.154250 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8blb2\" (UniqueName: \"kubernetes.io/projected/222fa6f9-b50a-4ef2-82fd-7b9d37841def-kube-api-access-8blb2\") pod \"nmstate-metrics-7f946cbc9-p442k\" (UID: \"222fa6f9-b50a-4ef2-82fd-7b9d37841def\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.154846 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8glrq\" (UniqueName: \"kubernetes.io/projected/6d436024-c16c-4428-81db-a144ed2deecc-kube-api-access-8glrq\") pod \"nmstate-webhook-5f6d4c5ccb-tcgqr\" (UID: \"6d436024-c16c-4428-81db-a144ed2deecc\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.173026 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.206001 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.207870 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5f5844bdbf-xdd6b"] Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.208646 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.226584 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f5844bdbf-xdd6b"] Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.229635 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/79eac56e-1d44-4f6b-875a-8f944ee63f7a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.229697 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sffbj\" (UniqueName: \"kubernetes.io/projected/79eac56e-1d44-4f6b-875a-8f944ee63f7a-kube-api-access-sffbj\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.229778 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/79eac56e-1d44-4f6b-875a-8f944ee63f7a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.230706 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/79eac56e-1d44-4f6b-875a-8f944ee63f7a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.234806 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.240409 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/79eac56e-1d44-4f6b-875a-8f944ee63f7a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.259210 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sffbj\" (UniqueName: \"kubernetes.io/projected/79eac56e-1d44-4f6b-875a-8f944ee63f7a-kube-api-access-sffbj\") pod \"nmstate-console-plugin-7fbb5f6569-fm2ks\" (UID: \"79eac56e-1d44-4f6b-875a-8f944ee63f7a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: W1205 12:19:58.272976 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ef69c53_a12b_4ae8_ae6f_baa299acfb6a.slice/crio-3552cdbb48e6facca402d8054a3248daabf8322f5f17b8c8734dc7cc5429c7e2 WatchSource:0}: Error finding container 3552cdbb48e6facca402d8054a3248daabf8322f5f17b8c8734dc7cc5429c7e2: Status 404 returned error can't find the container with id 3552cdbb48e6facca402d8054a3248daabf8322f5f17b8c8734dc7cc5429c7e2 Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332139 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332667 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-config\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332716 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-oauth-serving-cert\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332749 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-oauth-config\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332801 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-serving-cert\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332844 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhktl\" (UniqueName: \"kubernetes.io/projected/ccee4f97-4c08-4396-af2e-c30ccae24fc7-kube-api-access-nhktl\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332862 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-service-ca\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.332899 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-trusted-ca-bundle\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.408843 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p442k"] Dec 05 12:19:58 crc kubenswrapper[4807]: W1205 12:19:58.425448 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod222fa6f9_b50a_4ef2_82fd_7b9d37841def.slice/crio-8399cfa9d12b843b57ea438395a2613fedf0133e492ea58507c75eccf88a55a7 WatchSource:0}: Error finding container 8399cfa9d12b843b57ea438395a2613fedf0133e492ea58507c75eccf88a55a7: Status 404 returned error can't find the container with id 8399cfa9d12b843b57ea438395a2613fedf0133e492ea58507c75eccf88a55a7 Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.433812 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-config\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.433851 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-oauth-serving-cert\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.433892 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-oauth-config\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.433940 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-serving-cert\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.433988 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhktl\" (UniqueName: \"kubernetes.io/projected/ccee4f97-4c08-4396-af2e-c30ccae24fc7-kube-api-access-nhktl\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.434004 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-service-ca\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.434030 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-trusted-ca-bundle\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.435227 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-config\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.435562 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-trusted-ca-bundle\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.437228 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-oauth-serving-cert\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.438882 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-service-ca\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.441552 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-oauth-config\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.443480 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-serving-cert\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:58 crc kubenswrapper[4807]: I1205 12:19:58.455312 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhktl\" (UniqueName: \"kubernetes.io/projected/ccee4f97-4c08-4396-af2e-c30ccae24fc7-kube-api-access-nhktl\") pod \"console-5f5844bdbf-xdd6b\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:59 crc kubenswrapper[4807]: I1205 12:19:59.140294 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:19:59 crc kubenswrapper[4807]: I1205 12:19:59.166659 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hp4qg" event={"ID":"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a","Type":"ContainerStarted","Data":"3552cdbb48e6facca402d8054a3248daabf8322f5f17b8c8734dc7cc5429c7e2"} Dec 05 12:19:59 crc kubenswrapper[4807]: I1205 12:19:59.182866 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" event={"ID":"222fa6f9-b50a-4ef2-82fd-7b9d37841def","Type":"ContainerStarted","Data":"8399cfa9d12b843b57ea438395a2613fedf0133e492ea58507c75eccf88a55a7"} Dec 05 12:19:59 crc kubenswrapper[4807]: I1205 12:19:59.207102 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr"] Dec 05 12:19:59 crc kubenswrapper[4807]: I1205 12:19:59.214427 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks"] Dec 05 12:19:59 crc kubenswrapper[4807]: W1205 12:19:59.245762 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79eac56e_1d44_4f6b_875a_8f944ee63f7a.slice/crio-d9515aba2f5ac48869e38ce35ae8f9b1ab35c5583aa52452eb65283114fae66c WatchSource:0}: Error finding container d9515aba2f5ac48869e38ce35ae8f9b1ab35c5583aa52452eb65283114fae66c: Status 404 returned error can't find the container with id d9515aba2f5ac48869e38ce35ae8f9b1ab35c5583aa52452eb65283114fae66c Dec 05 12:19:59 crc kubenswrapper[4807]: I1205 12:19:59.385798 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f5844bdbf-xdd6b"] Dec 05 12:19:59 crc kubenswrapper[4807]: W1205 12:19:59.392786 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccee4f97_4c08_4396_af2e_c30ccae24fc7.slice/crio-2cc755b8d8ef52537ab7122546a5f607f146c0628e901841418f9ce986d6435e WatchSource:0}: Error finding container 2cc755b8d8ef52537ab7122546a5f607f146c0628e901841418f9ce986d6435e: Status 404 returned error can't find the container with id 2cc755b8d8ef52537ab7122546a5f607f146c0628e901841418f9ce986d6435e Dec 05 12:20:00 crc kubenswrapper[4807]: I1205 12:20:00.189443 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" event={"ID":"6d436024-c16c-4428-81db-a144ed2deecc","Type":"ContainerStarted","Data":"2034da7865b87e514b1dec2488ca5e018334cfab54a420fae7a57627b8ab28c5"} Dec 05 12:20:00 crc kubenswrapper[4807]: I1205 12:20:00.190361 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f5844bdbf-xdd6b" event={"ID":"ccee4f97-4c08-4396-af2e-c30ccae24fc7","Type":"ContainerStarted","Data":"2cc755b8d8ef52537ab7122546a5f607f146c0628e901841418f9ce986d6435e"} Dec 05 12:20:00 crc kubenswrapper[4807]: I1205 12:20:00.191165 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" event={"ID":"79eac56e-1d44-4f6b-875a-8f944ee63f7a","Type":"ContainerStarted","Data":"d9515aba2f5ac48869e38ce35ae8f9b1ab35c5583aa52452eb65283114fae66c"} Dec 05 12:20:03 crc kubenswrapper[4807]: I1205 12:20:03.206046 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f5844bdbf-xdd6b" event={"ID":"ccee4f97-4c08-4396-af2e-c30ccae24fc7","Type":"ContainerStarted","Data":"09483f27b026237683a2014ed52c9ff20960fa7cdf81da543af2169f92e22dc0"} Dec 05 12:20:03 crc kubenswrapper[4807]: I1205 12:20:03.227148 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5f5844bdbf-xdd6b" podStartSLOduration=5.227135077 podStartE2EDuration="5.227135077s" podCreationTimestamp="2025-12-05 12:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:20:03.223403324 +0000 UTC m=+832.717266593" watchObservedRunningTime="2025-12-05 12:20:03.227135077 +0000 UTC m=+832.720998346" Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.233410 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" event={"ID":"79eac56e-1d44-4f6b-875a-8f944ee63f7a","Type":"ContainerStarted","Data":"ec916fc955279a4788969408ee3328a546f0b7124e9057598137b5de2fdc8391"} Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.245954 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" event={"ID":"6d436024-c16c-4428-81db-a144ed2deecc","Type":"ContainerStarted","Data":"e74de996f9568282a2516fd9defafa66c4d5ab42a405ec68d20a0165fc962742"} Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.246017 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hp4qg" event={"ID":"1ef69c53-a12b-4ae8-ae6f-baa299acfb6a","Type":"ContainerStarted","Data":"255c1edc619159d5b4d46b41785b45ab1a971a6b539c2ac2f58fc1e36b2ba1b8"} Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.246033 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" event={"ID":"222fa6f9-b50a-4ef2-82fd-7b9d37841def","Type":"ContainerStarted","Data":"f00850988c4ec59c4cc5782f71a1e3afc906b594e7fcf48551ea7ee2081ec4f8"} Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.246055 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.246070 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.259320 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-fm2ks" podStartSLOduration=3.253525323 podStartE2EDuration="10.259304899s" podCreationTimestamp="2025-12-05 12:19:57 +0000 UTC" firstStartedPulling="2025-12-05 12:19:59.250084162 +0000 UTC m=+828.743947431" lastFinishedPulling="2025-12-05 12:20:06.255863698 +0000 UTC m=+835.749727007" observedRunningTime="2025-12-05 12:20:07.258111469 +0000 UTC m=+836.751974738" watchObservedRunningTime="2025-12-05 12:20:07.259304899 +0000 UTC m=+836.753168168" Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.276600 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" podStartSLOduration=3.245541253 podStartE2EDuration="10.27657836s" podCreationTimestamp="2025-12-05 12:19:57 +0000 UTC" firstStartedPulling="2025-12-05 12:19:59.249677732 +0000 UTC m=+828.743541001" lastFinishedPulling="2025-12-05 12:20:06.280714819 +0000 UTC m=+835.774578108" observedRunningTime="2025-12-05 12:20:07.274832196 +0000 UTC m=+836.768695465" watchObservedRunningTime="2025-12-05 12:20:07.27657836 +0000 UTC m=+836.770441629" Dec 05 12:20:07 crc kubenswrapper[4807]: I1205 12:20:07.298047 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-hp4qg" podStartSLOduration=2.273887367 podStartE2EDuration="10.298030676s" podCreationTimestamp="2025-12-05 12:19:57 +0000 UTC" firstStartedPulling="2025-12-05 12:19:58.276064386 +0000 UTC m=+827.769927645" lastFinishedPulling="2025-12-05 12:20:06.300207675 +0000 UTC m=+835.794070954" observedRunningTime="2025-12-05 12:20:07.294374974 +0000 UTC m=+836.788238243" watchObservedRunningTime="2025-12-05 12:20:07.298030676 +0000 UTC m=+836.791893945" Dec 05 12:20:09 crc kubenswrapper[4807]: I1205 12:20:09.141568 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:20:09 crc kubenswrapper[4807]: I1205 12:20:09.141611 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:20:09 crc kubenswrapper[4807]: I1205 12:20:09.146077 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:20:09 crc kubenswrapper[4807]: I1205 12:20:09.256820 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" event={"ID":"222fa6f9-b50a-4ef2-82fd-7b9d37841def","Type":"ContainerStarted","Data":"3ab160ccec0e813945b2e15fc627880459792cf83d66b371682ca1defab4b70b"} Dec 05 12:20:09 crc kubenswrapper[4807]: I1205 12:20:09.261387 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:20:09 crc kubenswrapper[4807]: I1205 12:20:09.273513 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p442k" podStartSLOduration=1.6412151019999999 podStartE2EDuration="12.273497793s" podCreationTimestamp="2025-12-05 12:19:57 +0000 UTC" firstStartedPulling="2025-12-05 12:19:58.43284888 +0000 UTC m=+827.926712139" lastFinishedPulling="2025-12-05 12:20:09.065131561 +0000 UTC m=+838.558994830" observedRunningTime="2025-12-05 12:20:09.271100423 +0000 UTC m=+838.764963692" watchObservedRunningTime="2025-12-05 12:20:09.273497793 +0000 UTC m=+838.767361062" Dec 05 12:20:09 crc kubenswrapper[4807]: I1205 12:20:09.340625 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kf4mv"] Dec 05 12:20:13 crc kubenswrapper[4807]: I1205 12:20:13.264903 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-hp4qg" Dec 05 12:20:18 crc kubenswrapper[4807]: I1205 12:20:18.213921 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-tcgqr" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.438478 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8"] Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.440185 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.445839 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.447242 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8"] Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.466477 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.466674 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.466755 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kdg6\" (UniqueName: \"kubernetes.io/projected/246b60bf-03b5-4139-a13d-691da402cc5f-kube-api-access-7kdg6\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.568448 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.569075 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.569116 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kdg6\" (UniqueName: \"kubernetes.io/projected/246b60bf-03b5-4139-a13d-691da402cc5f-kube-api-access-7kdg6\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.569437 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.569437 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.588395 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kdg6\" (UniqueName: \"kubernetes.io/projected/246b60bf-03b5-4139-a13d-691da402cc5f-kube-api-access-7kdg6\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:31 crc kubenswrapper[4807]: I1205 12:20:31.800319 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:32 crc kubenswrapper[4807]: I1205 12:20:32.214489 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8"] Dec 05 12:20:32 crc kubenswrapper[4807]: I1205 12:20:32.385953 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" event={"ID":"246b60bf-03b5-4139-a13d-691da402cc5f","Type":"ContainerStarted","Data":"fa515070c540de2c8c1e7acc4100e414c123a0cb5c1329c64bc0c4fb16bd1063"} Dec 05 12:20:32 crc kubenswrapper[4807]: I1205 12:20:32.386032 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" event={"ID":"246b60bf-03b5-4139-a13d-691da402cc5f","Type":"ContainerStarted","Data":"21bdbfc07e12cce42f21efefad3ff605a49b7acd6065c54305cc482b781f1926"} Dec 05 12:20:33 crc kubenswrapper[4807]: I1205 12:20:33.393469 4807 generic.go:334] "Generic (PLEG): container finished" podID="246b60bf-03b5-4139-a13d-691da402cc5f" containerID="fa515070c540de2c8c1e7acc4100e414c123a0cb5c1329c64bc0c4fb16bd1063" exitCode=0 Dec 05 12:20:33 crc kubenswrapper[4807]: I1205 12:20:33.393563 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" event={"ID":"246b60bf-03b5-4139-a13d-691da402cc5f","Type":"ContainerDied","Data":"fa515070c540de2c8c1e7acc4100e414c123a0cb5c1329c64bc0c4fb16bd1063"} Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.387411 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-kf4mv" podUID="7bbff2d6-47e4-489b-9c99-b3ef91531447" containerName="console" containerID="cri-o://b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280" gracePeriod=15 Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.782309 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kf4mv_7bbff2d6-47e4-489b-9c99-b3ef91531447/console/0.log" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.782557 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.806270 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-serving-cert\") pod \"7bbff2d6-47e4-489b-9c99-b3ef91531447\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.806321 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-service-ca\") pod \"7bbff2d6-47e4-489b-9c99-b3ef91531447\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.806357 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-trusted-ca-bundle\") pod \"7bbff2d6-47e4-489b-9c99-b3ef91531447\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.806388 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-oauth-config\") pod \"7bbff2d6-47e4-489b-9c99-b3ef91531447\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.806425 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vplzj\" (UniqueName: \"kubernetes.io/projected/7bbff2d6-47e4-489b-9c99-b3ef91531447-kube-api-access-vplzj\") pod \"7bbff2d6-47e4-489b-9c99-b3ef91531447\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.806447 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-config\") pod \"7bbff2d6-47e4-489b-9c99-b3ef91531447\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.806484 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-oauth-serving-cert\") pod \"7bbff2d6-47e4-489b-9c99-b3ef91531447\" (UID: \"7bbff2d6-47e4-489b-9c99-b3ef91531447\") " Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.807251 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7bbff2d6-47e4-489b-9c99-b3ef91531447" (UID: "7bbff2d6-47e4-489b-9c99-b3ef91531447"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.807287 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-service-ca" (OuterVolumeSpecName: "service-ca") pod "7bbff2d6-47e4-489b-9c99-b3ef91531447" (UID: "7bbff2d6-47e4-489b-9c99-b3ef91531447"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.807594 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-config" (OuterVolumeSpecName: "console-config") pod "7bbff2d6-47e4-489b-9c99-b3ef91531447" (UID: "7bbff2d6-47e4-489b-9c99-b3ef91531447"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.807756 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7bbff2d6-47e4-489b-9c99-b3ef91531447" (UID: "7bbff2d6-47e4-489b-9c99-b3ef91531447"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.814475 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7bbff2d6-47e4-489b-9c99-b3ef91531447" (UID: "7bbff2d6-47e4-489b-9c99-b3ef91531447"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.814665 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7bbff2d6-47e4-489b-9c99-b3ef91531447" (UID: "7bbff2d6-47e4-489b-9c99-b3ef91531447"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.815470 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bbff2d6-47e4-489b-9c99-b3ef91531447-kube-api-access-vplzj" (OuterVolumeSpecName: "kube-api-access-vplzj") pod "7bbff2d6-47e4-489b-9c99-b3ef91531447" (UID: "7bbff2d6-47e4-489b-9c99-b3ef91531447"). InnerVolumeSpecName "kube-api-access-vplzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.907446 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.907483 4807 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.907494 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vplzj\" (UniqueName: \"kubernetes.io/projected/7bbff2d6-47e4-489b-9c99-b3ef91531447-kube-api-access-vplzj\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.907504 4807 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.907512 4807 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.907537 4807 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bbff2d6-47e4-489b-9c99-b3ef91531447-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:34 crc kubenswrapper[4807]: I1205 12:20:34.907547 4807 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bbff2d6-47e4-489b-9c99-b3ef91531447-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.407211 4807 generic.go:334] "Generic (PLEG): container finished" podID="246b60bf-03b5-4139-a13d-691da402cc5f" containerID="3742b62111a6ac1373748b8dc5d43124fe93c567b91ccfb9da286c2fef18db2a" exitCode=0 Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.407279 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" event={"ID":"246b60bf-03b5-4139-a13d-691da402cc5f","Type":"ContainerDied","Data":"3742b62111a6ac1373748b8dc5d43124fe93c567b91ccfb9da286c2fef18db2a"} Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.409516 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kf4mv_7bbff2d6-47e4-489b-9c99-b3ef91531447/console/0.log" Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.409568 4807 generic.go:334] "Generic (PLEG): container finished" podID="7bbff2d6-47e4-489b-9c99-b3ef91531447" containerID="b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280" exitCode=2 Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.409588 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kf4mv" event={"ID":"7bbff2d6-47e4-489b-9c99-b3ef91531447","Type":"ContainerDied","Data":"b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280"} Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.409608 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kf4mv" event={"ID":"7bbff2d6-47e4-489b-9c99-b3ef91531447","Type":"ContainerDied","Data":"3f084501e323d5b722edefe5b3cb87c0fa4f2f3b2af53da08994fcdb46868569"} Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.409624 4807 scope.go:117] "RemoveContainer" containerID="b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280" Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.409637 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kf4mv" Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.429435 4807 scope.go:117] "RemoveContainer" containerID="b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280" Dec 05 12:20:35 crc kubenswrapper[4807]: E1205 12:20:35.429883 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280\": container with ID starting with b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280 not found: ID does not exist" containerID="b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280" Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.429916 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280"} err="failed to get container status \"b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280\": rpc error: code = NotFound desc = could not find container \"b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280\": container with ID starting with b6f0555ad7ff2d9e5749a2f14e93c3c0c0375632a04939f89d564d4111fc3280 not found: ID does not exist" Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.439745 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kf4mv"] Dec 05 12:20:35 crc kubenswrapper[4807]: I1205 12:20:35.444004 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-kf4mv"] Dec 05 12:20:36 crc kubenswrapper[4807]: I1205 12:20:36.418095 4807 generic.go:334] "Generic (PLEG): container finished" podID="246b60bf-03b5-4139-a13d-691da402cc5f" containerID="3fac62339b47ad516fb04426cdb69dece356f85b53d189ccb592605ef2ce754b" exitCode=0 Dec 05 12:20:36 crc kubenswrapper[4807]: I1205 12:20:36.418151 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" event={"ID":"246b60bf-03b5-4139-a13d-691da402cc5f","Type":"ContainerDied","Data":"3fac62339b47ad516fb04426cdb69dece356f85b53d189ccb592605ef2ce754b"} Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.243227 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bbff2d6-47e4-489b-9c99-b3ef91531447" path="/var/lib/kubelet/pods/7bbff2d6-47e4-489b-9c99-b3ef91531447/volumes" Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.630874 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.743503 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-util\") pod \"246b60bf-03b5-4139-a13d-691da402cc5f\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.743580 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kdg6\" (UniqueName: \"kubernetes.io/projected/246b60bf-03b5-4139-a13d-691da402cc5f-kube-api-access-7kdg6\") pod \"246b60bf-03b5-4139-a13d-691da402cc5f\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.743625 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-bundle\") pod \"246b60bf-03b5-4139-a13d-691da402cc5f\" (UID: \"246b60bf-03b5-4139-a13d-691da402cc5f\") " Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.744566 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-bundle" (OuterVolumeSpecName: "bundle") pod "246b60bf-03b5-4139-a13d-691da402cc5f" (UID: "246b60bf-03b5-4139-a13d-691da402cc5f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.748574 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/246b60bf-03b5-4139-a13d-691da402cc5f-kube-api-access-7kdg6" (OuterVolumeSpecName: "kube-api-access-7kdg6") pod "246b60bf-03b5-4139-a13d-691da402cc5f" (UID: "246b60bf-03b5-4139-a13d-691da402cc5f"). InnerVolumeSpecName "kube-api-access-7kdg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.837345 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-util" (OuterVolumeSpecName: "util") pod "246b60bf-03b5-4139-a13d-691da402cc5f" (UID: "246b60bf-03b5-4139-a13d-691da402cc5f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.845278 4807 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.845346 4807 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/246b60bf-03b5-4139-a13d-691da402cc5f-util\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:37 crc kubenswrapper[4807]: I1205 12:20:37.845360 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kdg6\" (UniqueName: \"kubernetes.io/projected/246b60bf-03b5-4139-a13d-691da402cc5f-kube-api-access-7kdg6\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:38 crc kubenswrapper[4807]: I1205 12:20:38.433733 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" event={"ID":"246b60bf-03b5-4139-a13d-691da402cc5f","Type":"ContainerDied","Data":"21bdbfc07e12cce42f21efefad3ff605a49b7acd6065c54305cc482b781f1926"} Dec 05 12:20:38 crc kubenswrapper[4807]: I1205 12:20:38.433771 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21bdbfc07e12cce42f21efefad3ff605a49b7acd6065c54305cc482b781f1926" Dec 05 12:20:38 crc kubenswrapper[4807]: I1205 12:20:38.433846 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.797168 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9vc7z"] Dec 05 12:20:42 crc kubenswrapper[4807]: E1205 12:20:42.799046 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246b60bf-03b5-4139-a13d-691da402cc5f" containerName="pull" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.799167 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="246b60bf-03b5-4139-a13d-691da402cc5f" containerName="pull" Dec 05 12:20:42 crc kubenswrapper[4807]: E1205 12:20:42.799243 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bbff2d6-47e4-489b-9c99-b3ef91531447" containerName="console" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.799308 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bbff2d6-47e4-489b-9c99-b3ef91531447" containerName="console" Dec 05 12:20:42 crc kubenswrapper[4807]: E1205 12:20:42.799389 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246b60bf-03b5-4139-a13d-691da402cc5f" containerName="util" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.799449 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="246b60bf-03b5-4139-a13d-691da402cc5f" containerName="util" Dec 05 12:20:42 crc kubenswrapper[4807]: E1205 12:20:42.799542 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246b60bf-03b5-4139-a13d-691da402cc5f" containerName="extract" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.799608 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="246b60bf-03b5-4139-a13d-691da402cc5f" containerName="extract" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.799804 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bbff2d6-47e4-489b-9c99-b3ef91531447" containerName="console" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.799885 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="246b60bf-03b5-4139-a13d-691da402cc5f" containerName="extract" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.800976 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.805291 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crbdc\" (UniqueName: \"kubernetes.io/projected/39609174-d530-43e8-af9b-2bdeca685fb5-kube-api-access-crbdc\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.805397 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-utilities\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.805450 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-catalog-content\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.812160 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vc7z"] Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.906594 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crbdc\" (UniqueName: \"kubernetes.io/projected/39609174-d530-43e8-af9b-2bdeca685fb5-kube-api-access-crbdc\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.906872 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-utilities\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.907002 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-catalog-content\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.907423 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-utilities\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.907466 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-catalog-content\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:42 crc kubenswrapper[4807]: I1205 12:20:42.933026 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crbdc\" (UniqueName: \"kubernetes.io/projected/39609174-d530-43e8-af9b-2bdeca685fb5-kube-api-access-crbdc\") pod \"redhat-marketplace-9vc7z\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:43 crc kubenswrapper[4807]: I1205 12:20:43.117285 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:43 crc kubenswrapper[4807]: I1205 12:20:43.522401 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vc7z"] Dec 05 12:20:44 crc kubenswrapper[4807]: I1205 12:20:44.466140 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vc7z" event={"ID":"39609174-d530-43e8-af9b-2bdeca685fb5","Type":"ContainerStarted","Data":"7d918066b64382efead5d91fd74f054a7c3153b7bb10e4c1fe5b82fe7105cbbe"} Dec 05 12:20:45 crc kubenswrapper[4807]: I1205 12:20:45.474444 4807 generic.go:334] "Generic (PLEG): container finished" podID="39609174-d530-43e8-af9b-2bdeca685fb5" containerID="91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2" exitCode=0 Dec 05 12:20:45 crc kubenswrapper[4807]: I1205 12:20:45.474488 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vc7z" event={"ID":"39609174-d530-43e8-af9b-2bdeca685fb5","Type":"ContainerDied","Data":"91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2"} Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.484158 4807 generic.go:334] "Generic (PLEG): container finished" podID="39609174-d530-43e8-af9b-2bdeca685fb5" containerID="22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6" exitCode=0 Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.484203 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vc7z" event={"ID":"39609174-d530-43e8-af9b-2bdeca685fb5","Type":"ContainerDied","Data":"22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6"} Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.784254 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw"] Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.784998 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.787468 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.787483 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.787504 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.788455 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.788495 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7mc9d" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.817104 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw"] Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.961433 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x75t8\" (UniqueName: \"kubernetes.io/projected/f973bc90-75ad-4998-bbf1-50910dc9a636-kube-api-access-x75t8\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.961488 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f973bc90-75ad-4998-bbf1-50910dc9a636-apiservice-cert\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:46 crc kubenswrapper[4807]: I1205 12:20:46.961646 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f973bc90-75ad-4998-bbf1-50910dc9a636-webhook-cert\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.052302 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c"] Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.053195 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.054936 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.058546 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-xbxxq" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.058548 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.062782 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f973bc90-75ad-4998-bbf1-50910dc9a636-apiservice-cert\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.062925 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f973bc90-75ad-4998-bbf1-50910dc9a636-webhook-cert\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.062964 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x75t8\" (UniqueName: \"kubernetes.io/projected/f973bc90-75ad-4998-bbf1-50910dc9a636-kube-api-access-x75t8\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.068783 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f973bc90-75ad-4998-bbf1-50910dc9a636-webhook-cert\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.070561 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f973bc90-75ad-4998-bbf1-50910dc9a636-apiservice-cert\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.077821 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c"] Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.089623 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x75t8\" (UniqueName: \"kubernetes.io/projected/f973bc90-75ad-4998-bbf1-50910dc9a636-kube-api-access-x75t8\") pod \"metallb-operator-controller-manager-d9c5fbb8c-crqlw\" (UID: \"f973bc90-75ad-4998-bbf1-50910dc9a636\") " pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.099295 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.164199 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/64233fc3-af41-47db-8676-ba182b32e5f7-apiservice-cert\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.164892 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxzdw\" (UniqueName: \"kubernetes.io/projected/64233fc3-af41-47db-8676-ba182b32e5f7-kube-api-access-jxzdw\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.164956 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/64233fc3-af41-47db-8676-ba182b32e5f7-webhook-cert\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.265777 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxzdw\" (UniqueName: \"kubernetes.io/projected/64233fc3-af41-47db-8676-ba182b32e5f7-kube-api-access-jxzdw\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.265824 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/64233fc3-af41-47db-8676-ba182b32e5f7-webhook-cert\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.265863 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/64233fc3-af41-47db-8676-ba182b32e5f7-apiservice-cert\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.271071 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/64233fc3-af41-47db-8676-ba182b32e5f7-webhook-cert\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.288254 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/64233fc3-af41-47db-8676-ba182b32e5f7-apiservice-cert\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.300807 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxzdw\" (UniqueName: \"kubernetes.io/projected/64233fc3-af41-47db-8676-ba182b32e5f7-kube-api-access-jxzdw\") pod \"metallb-operator-webhook-server-745b94f7cd-2h78c\" (UID: \"64233fc3-af41-47db-8676-ba182b32e5f7\") " pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.342479 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw"] Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.371085 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.494978 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vc7z" event={"ID":"39609174-d530-43e8-af9b-2bdeca685fb5","Type":"ContainerStarted","Data":"eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391"} Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.497471 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" event={"ID":"f973bc90-75ad-4998-bbf1-50910dc9a636","Type":"ContainerStarted","Data":"b17c840dc7093b19ec69d041a2861a50c28095aab273d7e3be8c9ae5d247c666"} Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.521186 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9vc7z" podStartSLOduration=4.068792272 podStartE2EDuration="5.521165683s" podCreationTimestamp="2025-12-05 12:20:42 +0000 UTC" firstStartedPulling="2025-12-05 12:20:45.476458945 +0000 UTC m=+874.970322214" lastFinishedPulling="2025-12-05 12:20:46.928832356 +0000 UTC m=+876.422695625" observedRunningTime="2025-12-05 12:20:47.516243071 +0000 UTC m=+877.010106340" watchObservedRunningTime="2025-12-05 12:20:47.521165683 +0000 UTC m=+877.015028952" Dec 05 12:20:47 crc kubenswrapper[4807]: I1205 12:20:47.619819 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c"] Dec 05 12:20:47 crc kubenswrapper[4807]: W1205 12:20:47.625854 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64233fc3_af41_47db_8676_ba182b32e5f7.slice/crio-3bd8cad05398ec49a2584427c8aadd0df0fd8d433fe6ecc17a8221d45d9764e9 WatchSource:0}: Error finding container 3bd8cad05398ec49a2584427c8aadd0df0fd8d433fe6ecc17a8221d45d9764e9: Status 404 returned error can't find the container with id 3bd8cad05398ec49a2584427c8aadd0df0fd8d433fe6ecc17a8221d45d9764e9 Dec 05 12:20:48 crc kubenswrapper[4807]: I1205 12:20:48.504241 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" event={"ID":"64233fc3-af41-47db-8676-ba182b32e5f7","Type":"ContainerStarted","Data":"3bd8cad05398ec49a2584427c8aadd0df0fd8d433fe6ecc17a8221d45d9764e9"} Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.117845 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.118517 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.159006 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.542818 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" event={"ID":"f973bc90-75ad-4998-bbf1-50910dc9a636","Type":"ContainerStarted","Data":"4fd8740bf38011681c9b26307f5e37b0d419b07d7ce158e75bffbb51a6e7f131"} Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.542993 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.545099 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" event={"ID":"64233fc3-af41-47db-8676-ba182b32e5f7","Type":"ContainerStarted","Data":"8eea39173847b7a618fc4d142dc353671c7765fe7dec2017f7e56dab6c6fc93c"} Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.603246 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" podStartSLOduration=4.048225862 podStartE2EDuration="7.603222018s" podCreationTimestamp="2025-12-05 12:20:46 +0000 UTC" firstStartedPulling="2025-12-05 12:20:47.362986489 +0000 UTC m=+876.856849758" lastFinishedPulling="2025-12-05 12:20:50.917982645 +0000 UTC m=+880.411845914" observedRunningTime="2025-12-05 12:20:53.596390479 +0000 UTC m=+883.090253758" watchObservedRunningTime="2025-12-05 12:20:53.603222018 +0000 UTC m=+883.097085287" Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.626926 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.628938 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" podStartSLOduration=1.3622099460000001 podStartE2EDuration="6.628918052s" podCreationTimestamp="2025-12-05 12:20:47 +0000 UTC" firstStartedPulling="2025-12-05 12:20:47.629517386 +0000 UTC m=+877.123380655" lastFinishedPulling="2025-12-05 12:20:52.896225492 +0000 UTC m=+882.390088761" observedRunningTime="2025-12-05 12:20:53.62684923 +0000 UTC m=+883.120712499" watchObservedRunningTime="2025-12-05 12:20:53.628918052 +0000 UTC m=+883.122781331" Dec 05 12:20:53 crc kubenswrapper[4807]: I1205 12:20:53.985565 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vc7z"] Dec 05 12:20:54 crc kubenswrapper[4807]: I1205 12:20:54.551389 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:20:55 crc kubenswrapper[4807]: I1205 12:20:55.556044 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9vc7z" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="registry-server" containerID="cri-o://eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391" gracePeriod=2 Dec 05 12:20:55 crc kubenswrapper[4807]: I1205 12:20:55.935259 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.109885 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-utilities\") pod \"39609174-d530-43e8-af9b-2bdeca685fb5\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.110225 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-catalog-content\") pod \"39609174-d530-43e8-af9b-2bdeca685fb5\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.110278 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crbdc\" (UniqueName: \"kubernetes.io/projected/39609174-d530-43e8-af9b-2bdeca685fb5-kube-api-access-crbdc\") pod \"39609174-d530-43e8-af9b-2bdeca685fb5\" (UID: \"39609174-d530-43e8-af9b-2bdeca685fb5\") " Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.110917 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-utilities" (OuterVolumeSpecName: "utilities") pod "39609174-d530-43e8-af9b-2bdeca685fb5" (UID: "39609174-d530-43e8-af9b-2bdeca685fb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.116179 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39609174-d530-43e8-af9b-2bdeca685fb5-kube-api-access-crbdc" (OuterVolumeSpecName: "kube-api-access-crbdc") pod "39609174-d530-43e8-af9b-2bdeca685fb5" (UID: "39609174-d530-43e8-af9b-2bdeca685fb5"). InnerVolumeSpecName "kube-api-access-crbdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.131350 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39609174-d530-43e8-af9b-2bdeca685fb5" (UID: "39609174-d530-43e8-af9b-2bdeca685fb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.211232 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.211273 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39609174-d530-43e8-af9b-2bdeca685fb5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.211290 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crbdc\" (UniqueName: \"kubernetes.io/projected/39609174-d530-43e8-af9b-2bdeca685fb5-kube-api-access-crbdc\") on node \"crc\" DevicePath \"\"" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.564430 4807 generic.go:334] "Generic (PLEG): container finished" podID="39609174-d530-43e8-af9b-2bdeca685fb5" containerID="eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391" exitCode=0 Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.564484 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vc7z" event={"ID":"39609174-d530-43e8-af9b-2bdeca685fb5","Type":"ContainerDied","Data":"eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391"} Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.564490 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vc7z" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.564517 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vc7z" event={"ID":"39609174-d530-43e8-af9b-2bdeca685fb5","Type":"ContainerDied","Data":"7d918066b64382efead5d91fd74f054a7c3153b7bb10e4c1fe5b82fe7105cbbe"} Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.564569 4807 scope.go:117] "RemoveContainer" containerID="eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.584324 4807 scope.go:117] "RemoveContainer" containerID="22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.601124 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vc7z"] Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.614761 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vc7z"] Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.607398 4807 scope.go:117] "RemoveContainer" containerID="91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.628344 4807 scope.go:117] "RemoveContainer" containerID="eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391" Dec 05 12:20:56 crc kubenswrapper[4807]: E1205 12:20:56.628884 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391\": container with ID starting with eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391 not found: ID does not exist" containerID="eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.628948 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391"} err="failed to get container status \"eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391\": rpc error: code = NotFound desc = could not find container \"eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391\": container with ID starting with eb8a90dcd77d5a4d85e82b27ce6da2f1ee2917831879c3bd6fb4dee2e6441391 not found: ID does not exist" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.628988 4807 scope.go:117] "RemoveContainer" containerID="22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6" Dec 05 12:20:56 crc kubenswrapper[4807]: E1205 12:20:56.629348 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6\": container with ID starting with 22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6 not found: ID does not exist" containerID="22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.629375 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6"} err="failed to get container status \"22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6\": rpc error: code = NotFound desc = could not find container \"22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6\": container with ID starting with 22c39d141f2e5c999c1678e2182acd0b5df6970696bccceb217a340578a4e4a6 not found: ID does not exist" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.629398 4807 scope.go:117] "RemoveContainer" containerID="91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2" Dec 05 12:20:56 crc kubenswrapper[4807]: E1205 12:20:56.629656 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2\": container with ID starting with 91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2 not found: ID does not exist" containerID="91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2" Dec 05 12:20:56 crc kubenswrapper[4807]: I1205 12:20:56.629697 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2"} err="failed to get container status \"91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2\": rpc error: code = NotFound desc = could not find container \"91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2\": container with ID starting with 91e5fa6445632dfecd88f799010a062010495ddab9487f448540d6fbd34834b2 not found: ID does not exist" Dec 05 12:20:57 crc kubenswrapper[4807]: I1205 12:20:57.243414 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" path="/var/lib/kubelet/pods/39609174-d530-43e8-af9b-2bdeca685fb5/volumes" Dec 05 12:21:07 crc kubenswrapper[4807]: I1205 12:21:07.375892 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-745b94f7cd-2h78c" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.103604 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-d9c5fbb8c-crqlw" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.979109 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4"] Dec 05 12:21:27 crc kubenswrapper[4807]: E1205 12:21:27.979408 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="extract-utilities" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.979430 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="extract-utilities" Dec 05 12:21:27 crc kubenswrapper[4807]: E1205 12:21:27.979452 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="extract-content" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.979462 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="extract-content" Dec 05 12:21:27 crc kubenswrapper[4807]: E1205 12:21:27.979474 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="registry-server" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.979483 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="registry-server" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.979627 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="39609174-d530-43e8-af9b-2bdeca685fb5" containerName="registry-server" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.980030 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.992993 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6ppt7"] Dec 05 12:21:27 crc kubenswrapper[4807]: I1205 12:21:27.996654 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.000642 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.000908 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.001093 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dppdz" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.001094 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.016782 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4"] Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103618 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-reloader\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103663 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103714 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-sockets\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103838 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq4q7\" (UniqueName: \"kubernetes.io/projected/f7a1e998-d292-49df-88b2-832d0ccb1c83-kube-api-access-qq4q7\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103866 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-startup\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103939 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6wtk4\" (UID: \"b144d4af-48d9-455d-8e32-3f5dc82cb3cf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103965 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tjmk\" (UniqueName: \"kubernetes.io/projected/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-kube-api-access-4tjmk\") pod \"frr-k8s-webhook-server-7fcb986d4-6wtk4\" (UID: \"b144d4af-48d9-455d-8e32-3f5dc82cb3cf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.103988 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics-certs\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.104024 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-conf\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.134662 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5tvpp"] Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.135857 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.140641 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.140659 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.140661 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.141207 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-h567t" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.142548 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-52sxs"] Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.143671 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.146459 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.196499 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-52sxs"] Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.206048 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq4q7\" (UniqueName: \"kubernetes.io/projected/f7a1e998-d292-49df-88b2-832d0ccb1c83-kube-api-access-qq4q7\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.206429 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-startup\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.206475 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6wtk4\" (UID: \"b144d4af-48d9-455d-8e32-3f5dc82cb3cf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.214672 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tjmk\" (UniqueName: \"kubernetes.io/projected/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-kube-api-access-4tjmk\") pod \"frr-k8s-webhook-server-7fcb986d4-6wtk4\" (UID: \"b144d4af-48d9-455d-8e32-3f5dc82cb3cf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.214788 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics-certs\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.214827 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-conf\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.214870 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-reloader\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.214894 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.214887 4807 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.215026 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-cert podName:b144d4af-48d9-455d-8e32-3f5dc82cb3cf nodeName:}" failed. No retries permitted until 2025-12-05 12:21:28.714965463 +0000 UTC m=+918.208828802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-cert") pod "frr-k8s-webhook-server-7fcb986d4-6wtk4" (UID: "b144d4af-48d9-455d-8e32-3f5dc82cb3cf") : secret "frr-k8s-webhook-server-cert" not found Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.215060 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-sockets\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.215101 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-startup\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.215391 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.215497 4807 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.215622 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics-certs podName:f7a1e998-d292-49df-88b2-832d0ccb1c83 nodeName:}" failed. No retries permitted until 2025-12-05 12:21:28.715602288 +0000 UTC m=+918.209465657 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics-certs") pod "frr-k8s-6ppt7" (UID: "f7a1e998-d292-49df-88b2-832d0ccb1c83") : secret "frr-k8s-certs-secret" not found Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.216717 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-conf\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.217926 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-frr-sockets\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.220774 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f7a1e998-d292-49df-88b2-832d0ccb1c83-reloader\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.262484 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tjmk\" (UniqueName: \"kubernetes.io/projected/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-kube-api-access-4tjmk\") pod \"frr-k8s-webhook-server-7fcb986d4-6wtk4\" (UID: \"b144d4af-48d9-455d-8e32-3f5dc82cb3cf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.262485 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq4q7\" (UniqueName: \"kubernetes.io/projected/f7a1e998-d292-49df-88b2-832d0ccb1c83-kube-api-access-qq4q7\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.316657 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-metrics-certs\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.316714 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.316741 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wbtp\" (UniqueName: \"kubernetes.io/projected/8532b350-fb6d-442c-b388-9a840576a99b-kube-api-access-8wbtp\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.316790 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/26eb0077-d64d-4b66-90ee-9e994c94916f-metrics-certs\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.317229 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26eb0077-d64d-4b66-90ee-9e994c94916f-cert\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.317304 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shk9s\" (UniqueName: \"kubernetes.io/projected/26eb0077-d64d-4b66-90ee-9e994c94916f-kube-api-access-shk9s\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.317340 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8532b350-fb6d-442c-b388-9a840576a99b-metallb-excludel2\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.418631 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-metrics-certs\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.418677 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.418697 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wbtp\" (UniqueName: \"kubernetes.io/projected/8532b350-fb6d-442c-b388-9a840576a99b-kube-api-access-8wbtp\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.418733 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/26eb0077-d64d-4b66-90ee-9e994c94916f-metrics-certs\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.418756 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26eb0077-d64d-4b66-90ee-9e994c94916f-cert\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.418774 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shk9s\" (UniqueName: \"kubernetes.io/projected/26eb0077-d64d-4b66-90ee-9e994c94916f-kube-api-access-shk9s\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.418792 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8532b350-fb6d-442c-b388-9a840576a99b-metallb-excludel2\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.418790 4807 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.418878 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-metrics-certs podName:8532b350-fb6d-442c-b388-9a840576a99b nodeName:}" failed. No retries permitted until 2025-12-05 12:21:28.918863104 +0000 UTC m=+918.412726373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-metrics-certs") pod "speaker-5tvpp" (UID: "8532b350-fb6d-442c-b388-9a840576a99b") : secret "speaker-certs-secret" not found Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.419330 4807 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.419392 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist podName:8532b350-fb6d-442c-b388-9a840576a99b nodeName:}" failed. No retries permitted until 2025-12-05 12:21:28.919378747 +0000 UTC m=+918.413242016 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist") pod "speaker-5tvpp" (UID: "8532b350-fb6d-442c-b388-9a840576a99b") : secret "metallb-memberlist" not found Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.419664 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8532b350-fb6d-442c-b388-9a840576a99b-metallb-excludel2\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.431606 4807 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.431967 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/26eb0077-d64d-4b66-90ee-9e994c94916f-metrics-certs\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.435217 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26eb0077-d64d-4b66-90ee-9e994c94916f-cert\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.438162 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wbtp\" (UniqueName: \"kubernetes.io/projected/8532b350-fb6d-442c-b388-9a840576a99b-kube-api-access-8wbtp\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.441598 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shk9s\" (UniqueName: \"kubernetes.io/projected/26eb0077-d64d-4b66-90ee-9e994c94916f-kube-api-access-shk9s\") pod \"controller-f8648f98b-52sxs\" (UID: \"26eb0077-d64d-4b66-90ee-9e994c94916f\") " pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.465144 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.723473 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6wtk4\" (UID: \"b144d4af-48d9-455d-8e32-3f5dc82cb3cf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.723898 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics-certs\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.727862 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7a1e998-d292-49df-88b2-832d0ccb1c83-metrics-certs\") pod \"frr-k8s-6ppt7\" (UID: \"f7a1e998-d292-49df-88b2-832d0ccb1c83\") " pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.727886 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b144d4af-48d9-455d-8e32-3f5dc82cb3cf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-6wtk4\" (UID: \"b144d4af-48d9-455d-8e32-3f5dc82cb3cf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.866850 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-52sxs"] Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.895926 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.926142 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-metrics-certs\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.926227 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.926434 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.926567 4807 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 12:21:28 crc kubenswrapper[4807]: E1205 12:21:28.926637 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist podName:8532b350-fb6d-442c-b388-9a840576a99b nodeName:}" failed. No retries permitted until 2025-12-05 12:21:29.926616154 +0000 UTC m=+919.420479423 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist") pod "speaker-5tvpp" (UID: "8532b350-fb6d-442c-b388-9a840576a99b") : secret "metallb-memberlist" not found Dec 05 12:21:28 crc kubenswrapper[4807]: I1205 12:21:28.933337 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-metrics-certs\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.146328 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4"] Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.759541 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" event={"ID":"b144d4af-48d9-455d-8e32-3f5dc82cb3cf","Type":"ContainerStarted","Data":"86f4177d6da41574574cd632c9102965d97fe7a67de89f57c7956d96e3e27e2a"} Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.760325 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerStarted","Data":"f60ea973c64fa4d396806b5857ffcdad78025dd3eb50b5d2244ae9f2b90a8634"} Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.761668 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-52sxs" event={"ID":"26eb0077-d64d-4b66-90ee-9e994c94916f","Type":"ContainerStarted","Data":"2e070b7f3784b84528a477cbf1f9c161aed545424ff7eb0ef44a0698481bb5b9"} Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.761702 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-52sxs" event={"ID":"26eb0077-d64d-4b66-90ee-9e994c94916f","Type":"ContainerStarted","Data":"b3e779ad1c6d3cdc205f7d5fe581600b7141dd5a15544228919117d9d41a5c12"} Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.761716 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-52sxs" event={"ID":"26eb0077-d64d-4b66-90ee-9e994c94916f","Type":"ContainerStarted","Data":"8f94cbc2e5f44f7280d51cfcc196d27cbc6467515f7e40cd6af2780e48109401"} Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.761810 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.942279 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:29 crc kubenswrapper[4807]: I1205 12:21:29.959237 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8532b350-fb6d-442c-b388-9a840576a99b-memberlist\") pod \"speaker-5tvpp\" (UID: \"8532b350-fb6d-442c-b388-9a840576a99b\") " pod="metallb-system/speaker-5tvpp" Dec 05 12:21:30 crc kubenswrapper[4807]: I1205 12:21:30.253912 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5tvpp" Dec 05 12:21:30 crc kubenswrapper[4807]: W1205 12:21:30.292268 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8532b350_fb6d_442c_b388_9a840576a99b.slice/crio-8eca6a8d1ad689458b2d8e2ff0200117e24a874c4b2edce4d8ae62fbcb77f643 WatchSource:0}: Error finding container 8eca6a8d1ad689458b2d8e2ff0200117e24a874c4b2edce4d8ae62fbcb77f643: Status 404 returned error can't find the container with id 8eca6a8d1ad689458b2d8e2ff0200117e24a874c4b2edce4d8ae62fbcb77f643 Dec 05 12:21:30 crc kubenswrapper[4807]: I1205 12:21:30.775634 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5tvpp" event={"ID":"8532b350-fb6d-442c-b388-9a840576a99b","Type":"ContainerStarted","Data":"551d2fef0d0d3a5cb8eb4360e4d231438cff410f99aaea95be8cfcc7bbb0eb03"} Dec 05 12:21:30 crc kubenswrapper[4807]: I1205 12:21:30.775956 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5tvpp" event={"ID":"8532b350-fb6d-442c-b388-9a840576a99b","Type":"ContainerStarted","Data":"8eca6a8d1ad689458b2d8e2ff0200117e24a874c4b2edce4d8ae62fbcb77f643"} Dec 05 12:21:31 crc kubenswrapper[4807]: I1205 12:21:31.265507 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-52sxs" podStartSLOduration=3.265489289 podStartE2EDuration="3.265489289s" podCreationTimestamp="2025-12-05 12:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:21:29.790728827 +0000 UTC m=+919.284592096" watchObservedRunningTime="2025-12-05 12:21:31.265489289 +0000 UTC m=+920.759352558" Dec 05 12:21:31 crc kubenswrapper[4807]: I1205 12:21:31.790892 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5tvpp" event={"ID":"8532b350-fb6d-442c-b388-9a840576a99b","Type":"ContainerStarted","Data":"110886af1fbbae88e6c11a603d4debcd72e193199829459550b496109f542afa"} Dec 05 12:21:31 crc kubenswrapper[4807]: I1205 12:21:31.791389 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5tvpp" Dec 05 12:21:36 crc kubenswrapper[4807]: I1205 12:21:36.831905 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" event={"ID":"b144d4af-48d9-455d-8e32-3f5dc82cb3cf","Type":"ContainerStarted","Data":"7d16af3c30983b9d9957095fabb98fca23c77964495a16a2abe686cd2e72871d"} Dec 05 12:21:36 crc kubenswrapper[4807]: I1205 12:21:36.832434 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:36 crc kubenswrapper[4807]: I1205 12:21:36.833366 4807 generic.go:334] "Generic (PLEG): container finished" podID="f7a1e998-d292-49df-88b2-832d0ccb1c83" containerID="6a70f3ed03fdc1c1ebdf8be7204dcc3bd7b712f3d783a77ccdccedd8c31bc45e" exitCode=0 Dec 05 12:21:36 crc kubenswrapper[4807]: I1205 12:21:36.833402 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerDied","Data":"6a70f3ed03fdc1c1ebdf8be7204dcc3bd7b712f3d783a77ccdccedd8c31bc45e"} Dec 05 12:21:36 crc kubenswrapper[4807]: I1205 12:21:36.868699 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" podStartSLOduration=2.600120254 podStartE2EDuration="9.868680568s" podCreationTimestamp="2025-12-05 12:21:27 +0000 UTC" firstStartedPulling="2025-12-05 12:21:29.164197627 +0000 UTC m=+918.658060896" lastFinishedPulling="2025-12-05 12:21:36.432757941 +0000 UTC m=+925.926621210" observedRunningTime="2025-12-05 12:21:36.866206967 +0000 UTC m=+926.360070236" watchObservedRunningTime="2025-12-05 12:21:36.868680568 +0000 UTC m=+926.362543837" Dec 05 12:21:36 crc kubenswrapper[4807]: I1205 12:21:36.869145 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5tvpp" podStartSLOduration=8.86914009 podStartE2EDuration="8.86914009s" podCreationTimestamp="2025-12-05 12:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:21:31.813349639 +0000 UTC m=+921.307212928" watchObservedRunningTime="2025-12-05 12:21:36.86914009 +0000 UTC m=+926.363003359" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.395494 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l89mw"] Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.397289 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.409860 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l89mw"] Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.547941 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-utilities\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.548008 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-catalog-content\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.548030 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb7zf\" (UniqueName: \"kubernetes.io/projected/bfc434fa-b958-4536-9e26-cab9925ede56-kube-api-access-lb7zf\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.649372 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-utilities\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.649458 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-catalog-content\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.649484 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb7zf\" (UniqueName: \"kubernetes.io/projected/bfc434fa-b958-4536-9e26-cab9925ede56-kube-api-access-lb7zf\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.649965 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-utilities\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.650005 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-catalog-content\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.673553 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb7zf\" (UniqueName: \"kubernetes.io/projected/bfc434fa-b958-4536-9e26-cab9925ede56-kube-api-access-lb7zf\") pod \"community-operators-l89mw\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.715584 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.845298 4807 generic.go:334] "Generic (PLEG): container finished" podID="f7a1e998-d292-49df-88b2-832d0ccb1c83" containerID="cbcd3dacb9da751bbc944be7e21594ffb8c2af620c29471429b5053b03570451" exitCode=0 Dec 05 12:21:37 crc kubenswrapper[4807]: I1205 12:21:37.845395 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerDied","Data":"cbcd3dacb9da751bbc944be7e21594ffb8c2af620c29471429b5053b03570451"} Dec 05 12:21:38 crc kubenswrapper[4807]: I1205 12:21:38.277047 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l89mw"] Dec 05 12:21:38 crc kubenswrapper[4807]: I1205 12:21:38.852534 4807 generic.go:334] "Generic (PLEG): container finished" podID="f7a1e998-d292-49df-88b2-832d0ccb1c83" containerID="7191948cbb678c571ceae667055f7d6b4cfb66a8809359f9b8bde3ff2bf4fe37" exitCode=0 Dec 05 12:21:38 crc kubenswrapper[4807]: I1205 12:21:38.852622 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerDied","Data":"7191948cbb678c571ceae667055f7d6b4cfb66a8809359f9b8bde3ff2bf4fe37"} Dec 05 12:21:38 crc kubenswrapper[4807]: I1205 12:21:38.855593 4807 generic.go:334] "Generic (PLEG): container finished" podID="bfc434fa-b958-4536-9e26-cab9925ede56" containerID="46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3" exitCode=0 Dec 05 12:21:38 crc kubenswrapper[4807]: I1205 12:21:38.855630 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l89mw" event={"ID":"bfc434fa-b958-4536-9e26-cab9925ede56","Type":"ContainerDied","Data":"46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3"} Dec 05 12:21:38 crc kubenswrapper[4807]: I1205 12:21:38.855705 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l89mw" event={"ID":"bfc434fa-b958-4536-9e26-cab9925ede56","Type":"ContainerStarted","Data":"fa9f139dd307f9f0f75443e00587b7eabcb4b9a86216683f81d00ac425c39934"} Dec 05 12:21:39 crc kubenswrapper[4807]: I1205 12:21:39.874401 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerStarted","Data":"23151f75529f34c8f03e1666e4991e1d71c62d94e6ef3afc2dd15777e57de959"} Dec 05 12:21:39 crc kubenswrapper[4807]: I1205 12:21:39.875039 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerStarted","Data":"8bf53ca3bdf8b727d0864527ece851bbf49e75f35a408cff83fc3c668c6805a3"} Dec 05 12:21:39 crc kubenswrapper[4807]: I1205 12:21:39.875055 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerStarted","Data":"b99bf5e036a47696d4a90c07d635bff22d350cf4375de41279de63192442e511"} Dec 05 12:21:39 crc kubenswrapper[4807]: I1205 12:21:39.875067 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerStarted","Data":"277494cb58014d2c40a8c694fd774c68d854d65833c47d1b91d238bdb8f9637b"} Dec 05 12:21:39 crc kubenswrapper[4807]: I1205 12:21:39.875078 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerStarted","Data":"620d175f020e119c6172d2c0161548ba9aee22644441434c1aafe4d123369dfc"} Dec 05 12:21:39 crc kubenswrapper[4807]: I1205 12:21:39.882182 4807 generic.go:334] "Generic (PLEG): container finished" podID="bfc434fa-b958-4536-9e26-cab9925ede56" containerID="19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c" exitCode=0 Dec 05 12:21:39 crc kubenswrapper[4807]: I1205 12:21:39.882227 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l89mw" event={"ID":"bfc434fa-b958-4536-9e26-cab9925ede56","Type":"ContainerDied","Data":"19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c"} Dec 05 12:21:40 crc kubenswrapper[4807]: I1205 12:21:40.258735 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5tvpp" Dec 05 12:21:40 crc kubenswrapper[4807]: I1205 12:21:40.890135 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l89mw" event={"ID":"bfc434fa-b958-4536-9e26-cab9925ede56","Type":"ContainerStarted","Data":"82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f"} Dec 05 12:21:40 crc kubenswrapper[4807]: I1205 12:21:40.894241 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6ppt7" event={"ID":"f7a1e998-d292-49df-88b2-832d0ccb1c83","Type":"ContainerStarted","Data":"70c1dec3700437135384db40e1d6f825cc2b8c7129556f6a3d0d10901b99ff33"} Dec 05 12:21:40 crc kubenswrapper[4807]: I1205 12:21:40.894885 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:40 crc kubenswrapper[4807]: I1205 12:21:40.920052 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l89mw" podStartSLOduration=2.445728901 podStartE2EDuration="3.920022592s" podCreationTimestamp="2025-12-05 12:21:37 +0000 UTC" firstStartedPulling="2025-12-05 12:21:38.85768444 +0000 UTC m=+928.351547709" lastFinishedPulling="2025-12-05 12:21:40.331978131 +0000 UTC m=+929.825841400" observedRunningTime="2025-12-05 12:21:40.918614947 +0000 UTC m=+930.412478226" watchObservedRunningTime="2025-12-05 12:21:40.920022592 +0000 UTC m=+930.413885901" Dec 05 12:21:40 crc kubenswrapper[4807]: I1205 12:21:40.943576 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6ppt7" podStartSLOduration=6.607632107 podStartE2EDuration="13.943517032s" podCreationTimestamp="2025-12-05 12:21:27 +0000 UTC" firstStartedPulling="2025-12-05 12:21:29.080321087 +0000 UTC m=+918.574184356" lastFinishedPulling="2025-12-05 12:21:36.416206012 +0000 UTC m=+925.910069281" observedRunningTime="2025-12-05 12:21:40.940962269 +0000 UTC m=+930.434825538" watchObservedRunningTime="2025-12-05 12:21:40.943517032 +0000 UTC m=+930.437380301" Dec 05 12:21:42 crc kubenswrapper[4807]: I1205 12:21:42.841415 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff"] Dec 05 12:21:42 crc kubenswrapper[4807]: I1205 12:21:42.843068 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:42 crc kubenswrapper[4807]: I1205 12:21:42.846988 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 12:21:42 crc kubenswrapper[4807]: I1205 12:21:42.851073 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff"] Dec 05 12:21:42 crc kubenswrapper[4807]: I1205 12:21:42.936112 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:42 crc kubenswrapper[4807]: I1205 12:21:42.936209 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bssw\" (UniqueName: \"kubernetes.io/projected/eddd7b9a-1c09-4b02-bc51-67cb07107100-kube-api-access-8bssw\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:42 crc kubenswrapper[4807]: I1205 12:21:42.936264 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.037581 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bssw\" (UniqueName: \"kubernetes.io/projected/eddd7b9a-1c09-4b02-bc51-67cb07107100-kube-api-access-8bssw\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.037663 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.037761 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.038329 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.038344 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.056725 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bssw\" (UniqueName: \"kubernetes.io/projected/eddd7b9a-1c09-4b02-bc51-67cb07107100-kube-api-access-8bssw\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.159963 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.611976 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff"] Dec 05 12:21:43 crc kubenswrapper[4807]: W1205 12:21:43.619775 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeddd7b9a_1c09_4b02_bc51_67cb07107100.slice/crio-8c7f3ee6e38a9e59f66a8e930e7a069227f6a000a8d0238246f263890651d55d WatchSource:0}: Error finding container 8c7f3ee6e38a9e59f66a8e930e7a069227f6a000a8d0238246f263890651d55d: Status 404 returned error can't find the container with id 8c7f3ee6e38a9e59f66a8e930e7a069227f6a000a8d0238246f263890651d55d Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.921235 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" event={"ID":"eddd7b9a-1c09-4b02-bc51-67cb07107100","Type":"ContainerStarted","Data":"8c7f3ee6e38a9e59f66a8e930e7a069227f6a000a8d0238246f263890651d55d"} Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.926705 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:43 crc kubenswrapper[4807]: I1205 12:21:43.968409 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:44 crc kubenswrapper[4807]: I1205 12:21:44.929443 4807 generic.go:334] "Generic (PLEG): container finished" podID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerID="3251218d53d9afb1caff28d8fbb9377f91cf93da5a8b43faf8ae2225bbd42b84" exitCode=0 Dec 05 12:21:44 crc kubenswrapper[4807]: I1205 12:21:44.930081 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" event={"ID":"eddd7b9a-1c09-4b02-bc51-67cb07107100","Type":"ContainerDied","Data":"3251218d53d9afb1caff28d8fbb9377f91cf93da5a8b43faf8ae2225bbd42b84"} Dec 05 12:21:46 crc kubenswrapper[4807]: I1205 12:21:46.998974 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rcxxn"] Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.003627 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.008655 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rcxxn"] Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.088226 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-catalog-content\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.088300 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-utilities\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.088402 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7bzq\" (UniqueName: \"kubernetes.io/projected/f9590f52-46f1-4ab1-8933-f0ad618957d4-kube-api-access-k7bzq\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.190502 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-catalog-content\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.190591 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-utilities\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.190683 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7bzq\" (UniqueName: \"kubernetes.io/projected/f9590f52-46f1-4ab1-8933-f0ad618957d4-kube-api-access-k7bzq\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.191394 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-catalog-content\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.191484 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-utilities\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.214663 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7bzq\" (UniqueName: \"kubernetes.io/projected/f9590f52-46f1-4ab1-8933-f0ad618957d4-kube-api-access-k7bzq\") pod \"certified-operators-rcxxn\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.327190 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.716689 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.717052 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.766656 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:47 crc kubenswrapper[4807]: I1205 12:21:47.982578 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:48 crc kubenswrapper[4807]: I1205 12:21:48.469798 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-52sxs" Dec 05 12:21:48 crc kubenswrapper[4807]: I1205 12:21:48.900977 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-6wtk4" Dec 05 12:21:48 crc kubenswrapper[4807]: I1205 12:21:48.930041 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6ppt7" Dec 05 12:21:49 crc kubenswrapper[4807]: I1205 12:21:49.598711 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rcxxn"] Dec 05 12:21:49 crc kubenswrapper[4807]: I1205 12:21:49.966269 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerStarted","Data":"a1d935e2beba81c4cbed6ea9aa71f249d7f28cbb0136c42da03016c4d97970a0"} Dec 05 12:21:49 crc kubenswrapper[4807]: I1205 12:21:49.966456 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerStarted","Data":"c356e8cf7b21d589d1963720a1ee074125757f5f1c531b0644dac41e235fd605"} Dec 05 12:21:50 crc kubenswrapper[4807]: I1205 12:21:50.977103 4807 generic.go:334] "Generic (PLEG): container finished" podID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerID="a1d935e2beba81c4cbed6ea9aa71f249d7f28cbb0136c42da03016c4d97970a0" exitCode=0 Dec 05 12:21:50 crc kubenswrapper[4807]: I1205 12:21:50.977147 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerDied","Data":"a1d935e2beba81c4cbed6ea9aa71f249d7f28cbb0136c42da03016c4d97970a0"} Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.387503 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l89mw"] Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.388107 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l89mw" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="registry-server" containerID="cri-o://82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f" gracePeriod=2 Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.814402 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.960224 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-catalog-content\") pod \"bfc434fa-b958-4536-9e26-cab9925ede56\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.960620 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-utilities\") pod \"bfc434fa-b958-4536-9e26-cab9925ede56\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.960738 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb7zf\" (UniqueName: \"kubernetes.io/projected/bfc434fa-b958-4536-9e26-cab9925ede56-kube-api-access-lb7zf\") pod \"bfc434fa-b958-4536-9e26-cab9925ede56\" (UID: \"bfc434fa-b958-4536-9e26-cab9925ede56\") " Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.961464 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-utilities" (OuterVolumeSpecName: "utilities") pod "bfc434fa-b958-4536-9e26-cab9925ede56" (UID: "bfc434fa-b958-4536-9e26-cab9925ede56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.966092 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfc434fa-b958-4536-9e26-cab9925ede56-kube-api-access-lb7zf" (OuterVolumeSpecName: "kube-api-access-lb7zf") pod "bfc434fa-b958-4536-9e26-cab9925ede56" (UID: "bfc434fa-b958-4536-9e26-cab9925ede56"). InnerVolumeSpecName "kube-api-access-lb7zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.984542 4807 generic.go:334] "Generic (PLEG): container finished" podID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerID="0fce4ec5011ab045322f6cc098075d0571ec9b16f95bde35d23c224a2496920a" exitCode=0 Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.984615 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" event={"ID":"eddd7b9a-1c09-4b02-bc51-67cb07107100","Type":"ContainerDied","Data":"0fce4ec5011ab045322f6cc098075d0571ec9b16f95bde35d23c224a2496920a"} Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.988030 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerStarted","Data":"e7daacd844c14e6920a61b6295cb831a16f56461e9b7696f8a5837f49a0335a3"} Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.991193 4807 generic.go:334] "Generic (PLEG): container finished" podID="bfc434fa-b958-4536-9e26-cab9925ede56" containerID="82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f" exitCode=0 Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.991236 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l89mw" event={"ID":"bfc434fa-b958-4536-9e26-cab9925ede56","Type":"ContainerDied","Data":"82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f"} Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.991262 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l89mw" event={"ID":"bfc434fa-b958-4536-9e26-cab9925ede56","Type":"ContainerDied","Data":"fa9f139dd307f9f0f75443e00587b7eabcb4b9a86216683f81d00ac425c39934"} Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.991282 4807 scope.go:117] "RemoveContainer" containerID="82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f" Dec 05 12:21:51 crc kubenswrapper[4807]: I1205 12:21:51.991452 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l89mw" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.007580 4807 scope.go:117] "RemoveContainer" containerID="19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.013403 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfc434fa-b958-4536-9e26-cab9925ede56" (UID: "bfc434fa-b958-4536-9e26-cab9925ede56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.025762 4807 scope.go:117] "RemoveContainer" containerID="46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.040418 4807 scope.go:117] "RemoveContainer" containerID="82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f" Dec 05 12:21:52 crc kubenswrapper[4807]: E1205 12:21:52.040815 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f\": container with ID starting with 82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f not found: ID does not exist" containerID="82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.040847 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f"} err="failed to get container status \"82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f\": rpc error: code = NotFound desc = could not find container \"82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f\": container with ID starting with 82021123d2bfa75dc1bb24ad8ee9038a1b2665204fd00c2b5f203c76846e136f not found: ID does not exist" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.041554 4807 scope.go:117] "RemoveContainer" containerID="19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c" Dec 05 12:21:52 crc kubenswrapper[4807]: E1205 12:21:52.041878 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c\": container with ID starting with 19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c not found: ID does not exist" containerID="19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.041905 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c"} err="failed to get container status \"19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c\": rpc error: code = NotFound desc = could not find container \"19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c\": container with ID starting with 19f5bbf483a8994841cc2d4c4b451f6a130a08ed5727608a08628eb8da48e63c not found: ID does not exist" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.041919 4807 scope.go:117] "RemoveContainer" containerID="46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3" Dec 05 12:21:52 crc kubenswrapper[4807]: E1205 12:21:52.042298 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3\": container with ID starting with 46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3 not found: ID does not exist" containerID="46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.042359 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3"} err="failed to get container status \"46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3\": rpc error: code = NotFound desc = could not find container \"46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3\": container with ID starting with 46a1d123f47f2849f7c1649464c32225323d89560584b060c9daeb993927c0f3 not found: ID does not exist" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.062213 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.062236 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc434fa-b958-4536-9e26-cab9925ede56-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.062245 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb7zf\" (UniqueName: \"kubernetes.io/projected/bfc434fa-b958-4536-9e26-cab9925ede56-kube-api-access-lb7zf\") on node \"crc\" DevicePath \"\"" Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.353509 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l89mw"] Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.361087 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l89mw"] Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.466681 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:21:52 crc kubenswrapper[4807]: I1205 12:21:52.466753 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:21:53 crc kubenswrapper[4807]: I1205 12:21:53.001042 4807 generic.go:334] "Generic (PLEG): container finished" podID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerID="4e92ee9472079f1f86bd6c0056df3d56d2906d318875fdf49c90c133354e1006" exitCode=0 Dec 05 12:21:53 crc kubenswrapper[4807]: I1205 12:21:53.001148 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" event={"ID":"eddd7b9a-1c09-4b02-bc51-67cb07107100","Type":"ContainerDied","Data":"4e92ee9472079f1f86bd6c0056df3d56d2906d318875fdf49c90c133354e1006"} Dec 05 12:21:53 crc kubenswrapper[4807]: I1205 12:21:53.004065 4807 generic.go:334] "Generic (PLEG): container finished" podID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerID="e7daacd844c14e6920a61b6295cb831a16f56461e9b7696f8a5837f49a0335a3" exitCode=0 Dec 05 12:21:53 crc kubenswrapper[4807]: I1205 12:21:53.004109 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerDied","Data":"e7daacd844c14e6920a61b6295cb831a16f56461e9b7696f8a5837f49a0335a3"} Dec 05 12:21:53 crc kubenswrapper[4807]: I1205 12:21:53.243915 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" path="/var/lib/kubelet/pods/bfc434fa-b958-4536-9e26-cab9925ede56/volumes" Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.266864 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.390409 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bssw\" (UniqueName: \"kubernetes.io/projected/eddd7b9a-1c09-4b02-bc51-67cb07107100-kube-api-access-8bssw\") pod \"eddd7b9a-1c09-4b02-bc51-67cb07107100\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.390611 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-bundle\") pod \"eddd7b9a-1c09-4b02-bc51-67cb07107100\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.390640 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-util\") pod \"eddd7b9a-1c09-4b02-bc51-67cb07107100\" (UID: \"eddd7b9a-1c09-4b02-bc51-67cb07107100\") " Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.392667 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-bundle" (OuterVolumeSpecName: "bundle") pod "eddd7b9a-1c09-4b02-bc51-67cb07107100" (UID: "eddd7b9a-1c09-4b02-bc51-67cb07107100"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.396255 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eddd7b9a-1c09-4b02-bc51-67cb07107100-kube-api-access-8bssw" (OuterVolumeSpecName: "kube-api-access-8bssw") pod "eddd7b9a-1c09-4b02-bc51-67cb07107100" (UID: "eddd7b9a-1c09-4b02-bc51-67cb07107100"). InnerVolumeSpecName "kube-api-access-8bssw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.401779 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-util" (OuterVolumeSpecName: "util") pod "eddd7b9a-1c09-4b02-bc51-67cb07107100" (UID: "eddd7b9a-1c09-4b02-bc51-67cb07107100"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.491923 4807 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.491953 4807 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eddd7b9a-1c09-4b02-bc51-67cb07107100-util\") on node \"crc\" DevicePath \"\"" Dec 05 12:21:54 crc kubenswrapper[4807]: I1205 12:21:54.491964 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bssw\" (UniqueName: \"kubernetes.io/projected/eddd7b9a-1c09-4b02-bc51-67cb07107100-kube-api-access-8bssw\") on node \"crc\" DevicePath \"\"" Dec 05 12:21:55 crc kubenswrapper[4807]: I1205 12:21:55.016960 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerStarted","Data":"2af576ac298a2319b48fd4efbdf12700976c60960ae0f9bd5dd004202d2c62e2"} Dec 05 12:21:55 crc kubenswrapper[4807]: I1205 12:21:55.019318 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" event={"ID":"eddd7b9a-1c09-4b02-bc51-67cb07107100","Type":"ContainerDied","Data":"8c7f3ee6e38a9e59f66a8e930e7a069227f6a000a8d0238246f263890651d55d"} Dec 05 12:21:55 crc kubenswrapper[4807]: I1205 12:21:55.019346 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c7f3ee6e38a9e59f66a8e930e7a069227f6a000a8d0238246f263890651d55d" Dec 05 12:21:55 crc kubenswrapper[4807]: I1205 12:21:55.019475 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff" Dec 05 12:21:55 crc kubenswrapper[4807]: I1205 12:21:55.037595 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rcxxn" podStartSLOduration=6.041145595 podStartE2EDuration="9.037579887s" podCreationTimestamp="2025-12-05 12:21:46 +0000 UTC" firstStartedPulling="2025-12-05 12:21:51.040988854 +0000 UTC m=+940.534852123" lastFinishedPulling="2025-12-05 12:21:54.037423146 +0000 UTC m=+943.531286415" observedRunningTime="2025-12-05 12:21:55.032755838 +0000 UTC m=+944.526619117" watchObservedRunningTime="2025-12-05 12:21:55.037579887 +0000 UTC m=+944.531443156" Dec 05 12:21:57 crc kubenswrapper[4807]: I1205 12:21:57.327515 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:57 crc kubenswrapper[4807]: I1205 12:21:57.327884 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:21:57 crc kubenswrapper[4807]: I1205 12:21:57.368243 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.000579 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc"] Dec 05 12:22:02 crc kubenswrapper[4807]: E1205 12:22:02.001180 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerName="extract" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001195 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerName="extract" Dec 05 12:22:02 crc kubenswrapper[4807]: E1205 12:22:02.001214 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerName="pull" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001222 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerName="pull" Dec 05 12:22:02 crc kubenswrapper[4807]: E1205 12:22:02.001242 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="extract-utilities" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001251 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="extract-utilities" Dec 05 12:22:02 crc kubenswrapper[4807]: E1205 12:22:02.001262 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerName="util" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001269 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerName="util" Dec 05 12:22:02 crc kubenswrapper[4807]: E1205 12:22:02.001283 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="extract-content" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001290 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="extract-content" Dec 05 12:22:02 crc kubenswrapper[4807]: E1205 12:22:02.001304 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="registry-server" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001312 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="registry-server" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001474 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfc434fa-b958-4536-9e26-cab9925ede56" containerName="registry-server" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.001488 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="eddd7b9a-1c09-4b02-bc51-67cb07107100" containerName="extract" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.002052 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.004267 4807 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-2frbs" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.004845 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.005045 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.054107 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc"] Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.197283 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7vbk\" (UniqueName: \"kubernetes.io/projected/c328cc5f-9c16-4088-bcaf-5f374601ff29-kube-api-access-n7vbk\") pod \"cert-manager-operator-controller-manager-64cf6dff88-dsfqc\" (UID: \"c328cc5f-9c16-4088-bcaf-5f374601ff29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.197378 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c328cc5f-9c16-4088-bcaf-5f374601ff29-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-dsfqc\" (UID: \"c328cc5f-9c16-4088-bcaf-5f374601ff29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.298083 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7vbk\" (UniqueName: \"kubernetes.io/projected/c328cc5f-9c16-4088-bcaf-5f374601ff29-kube-api-access-n7vbk\") pod \"cert-manager-operator-controller-manager-64cf6dff88-dsfqc\" (UID: \"c328cc5f-9c16-4088-bcaf-5f374601ff29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.298169 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c328cc5f-9c16-4088-bcaf-5f374601ff29-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-dsfqc\" (UID: \"c328cc5f-9c16-4088-bcaf-5f374601ff29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.299088 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c328cc5f-9c16-4088-bcaf-5f374601ff29-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-dsfqc\" (UID: \"c328cc5f-9c16-4088-bcaf-5f374601ff29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.323719 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7vbk\" (UniqueName: \"kubernetes.io/projected/c328cc5f-9c16-4088-bcaf-5f374601ff29-kube-api-access-n7vbk\") pod \"cert-manager-operator-controller-manager-64cf6dff88-dsfqc\" (UID: \"c328cc5f-9c16-4088-bcaf-5f374601ff29\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:02 crc kubenswrapper[4807]: I1205 12:22:02.616890 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" Dec 05 12:22:03 crc kubenswrapper[4807]: I1205 12:22:03.055342 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc"] Dec 05 12:22:03 crc kubenswrapper[4807]: W1205 12:22:03.062124 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc328cc5f_9c16_4088_bcaf_5f374601ff29.slice/crio-2e282741f8acd3b4a475cd07af74622a49f3acd063458455a90efa34f5c8d908 WatchSource:0}: Error finding container 2e282741f8acd3b4a475cd07af74622a49f3acd063458455a90efa34f5c8d908: Status 404 returned error can't find the container with id 2e282741f8acd3b4a475cd07af74622a49f3acd063458455a90efa34f5c8d908 Dec 05 12:22:03 crc kubenswrapper[4807]: I1205 12:22:03.097377 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" event={"ID":"c328cc5f-9c16-4088-bcaf-5f374601ff29","Type":"ContainerStarted","Data":"2e282741f8acd3b4a475cd07af74622a49f3acd063458455a90efa34f5c8d908"} Dec 05 12:22:07 crc kubenswrapper[4807]: I1205 12:22:07.124402 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" event={"ID":"c328cc5f-9c16-4088-bcaf-5f374601ff29","Type":"ContainerStarted","Data":"06999299431e27481b3cb4ddadaeaa7c60fa744f391e5a9806488746876ec9d7"} Dec 05 12:22:07 crc kubenswrapper[4807]: I1205 12:22:07.152390 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-dsfqc" podStartSLOduration=2.452087018 podStartE2EDuration="6.15236478s" podCreationTimestamp="2025-12-05 12:22:01 +0000 UTC" firstStartedPulling="2025-12-05 12:22:03.06348487 +0000 UTC m=+952.557348139" lastFinishedPulling="2025-12-05 12:22:06.763762632 +0000 UTC m=+956.257625901" observedRunningTime="2025-12-05 12:22:07.144003374 +0000 UTC m=+956.637866663" watchObservedRunningTime="2025-12-05 12:22:07.15236478 +0000 UTC m=+956.646228049" Dec 05 12:22:07 crc kubenswrapper[4807]: I1205 12:22:07.372122 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:22:08 crc kubenswrapper[4807]: I1205 12:22:08.989609 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rcxxn"] Dec 05 12:22:08 crc kubenswrapper[4807]: I1205 12:22:08.990111 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rcxxn" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="registry-server" containerID="cri-o://2af576ac298a2319b48fd4efbdf12700976c60960ae0f9bd5dd004202d2c62e2" gracePeriod=2 Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.145984 4807 generic.go:334] "Generic (PLEG): container finished" podID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerID="2af576ac298a2319b48fd4efbdf12700976c60960ae0f9bd5dd004202d2c62e2" exitCode=0 Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.146036 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerDied","Data":"2af576ac298a2319b48fd4efbdf12700976c60960ae0f9bd5dd004202d2c62e2"} Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.477209 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.500965 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7bzq\" (UniqueName: \"kubernetes.io/projected/f9590f52-46f1-4ab1-8933-f0ad618957d4-kube-api-access-k7bzq\") pod \"f9590f52-46f1-4ab1-8933-f0ad618957d4\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.501085 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-utilities\") pod \"f9590f52-46f1-4ab1-8933-f0ad618957d4\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.501213 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-catalog-content\") pod \"f9590f52-46f1-4ab1-8933-f0ad618957d4\" (UID: \"f9590f52-46f1-4ab1-8933-f0ad618957d4\") " Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.502252 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-utilities" (OuterVolumeSpecName: "utilities") pod "f9590f52-46f1-4ab1-8933-f0ad618957d4" (UID: "f9590f52-46f1-4ab1-8933-f0ad618957d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.514375 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9590f52-46f1-4ab1-8933-f0ad618957d4-kube-api-access-k7bzq" (OuterVolumeSpecName: "kube-api-access-k7bzq") pod "f9590f52-46f1-4ab1-8933-f0ad618957d4" (UID: "f9590f52-46f1-4ab1-8933-f0ad618957d4"). InnerVolumeSpecName "kube-api-access-k7bzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.555487 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9590f52-46f1-4ab1-8933-f0ad618957d4" (UID: "f9590f52-46f1-4ab1-8933-f0ad618957d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.602565 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7bzq\" (UniqueName: \"kubernetes.io/projected/f9590f52-46f1-4ab1-8933-f0ad618957d4-kube-api-access-k7bzq\") on node \"crc\" DevicePath \"\"" Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.602605 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:22:09 crc kubenswrapper[4807]: I1205 12:22:09.602622 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9590f52-46f1-4ab1-8933-f0ad618957d4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:22:10 crc kubenswrapper[4807]: I1205 12:22:10.153029 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rcxxn" event={"ID":"f9590f52-46f1-4ab1-8933-f0ad618957d4","Type":"ContainerDied","Data":"c356e8cf7b21d589d1963720a1ee074125757f5f1c531b0644dac41e235fd605"} Dec 05 12:22:10 crc kubenswrapper[4807]: I1205 12:22:10.153084 4807 scope.go:117] "RemoveContainer" containerID="2af576ac298a2319b48fd4efbdf12700976c60960ae0f9bd5dd004202d2c62e2" Dec 05 12:22:10 crc kubenswrapper[4807]: I1205 12:22:10.153228 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rcxxn" Dec 05 12:22:10 crc kubenswrapper[4807]: I1205 12:22:10.177178 4807 scope.go:117] "RemoveContainer" containerID="e7daacd844c14e6920a61b6295cb831a16f56461e9b7696f8a5837f49a0335a3" Dec 05 12:22:10 crc kubenswrapper[4807]: I1205 12:22:10.188504 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rcxxn"] Dec 05 12:22:10 crc kubenswrapper[4807]: I1205 12:22:10.193165 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rcxxn"] Dec 05 12:22:10 crc kubenswrapper[4807]: I1205 12:22:10.202864 4807 scope.go:117] "RemoveContainer" containerID="a1d935e2beba81c4cbed6ea9aa71f249d7f28cbb0136c42da03016c4d97970a0" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.036661 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-65hxg"] Dec 05 12:22:11 crc kubenswrapper[4807]: E1205 12:22:11.037014 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="registry-server" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.037039 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="registry-server" Dec 05 12:22:11 crc kubenswrapper[4807]: E1205 12:22:11.037065 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="extract-content" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.037078 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="extract-content" Dec 05 12:22:11 crc kubenswrapper[4807]: E1205 12:22:11.037105 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="extract-utilities" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.037118 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="extract-utilities" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.037326 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" containerName="registry-server" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.038011 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.040438 4807 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hf9pp" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.041041 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.041350 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.087292 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-65hxg"] Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.120188 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsxfv\" (UniqueName: \"kubernetes.io/projected/82008b0e-6099-474a-8526-f531424944ff-kube-api-access-tsxfv\") pod \"cert-manager-webhook-f4fb5df64-65hxg\" (UID: \"82008b0e-6099-474a-8526-f531424944ff\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.120237 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/82008b0e-6099-474a-8526-f531424944ff-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-65hxg\" (UID: \"82008b0e-6099-474a-8526-f531424944ff\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.221393 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsxfv\" (UniqueName: \"kubernetes.io/projected/82008b0e-6099-474a-8526-f531424944ff-kube-api-access-tsxfv\") pod \"cert-manager-webhook-f4fb5df64-65hxg\" (UID: \"82008b0e-6099-474a-8526-f531424944ff\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.221468 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/82008b0e-6099-474a-8526-f531424944ff-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-65hxg\" (UID: \"82008b0e-6099-474a-8526-f531424944ff\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.234010 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.237591 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/82008b0e-6099-474a-8526-f531424944ff-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-65hxg\" (UID: \"82008b0e-6099-474a-8526-f531424944ff\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.241959 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9590f52-46f1-4ab1-8933-f0ad618957d4" path="/var/lib/kubelet/pods/f9590f52-46f1-4ab1-8933-f0ad618957d4/volumes" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.244412 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.257870 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsxfv\" (UniqueName: \"kubernetes.io/projected/82008b0e-6099-474a-8526-f531424944ff-kube-api-access-tsxfv\") pod \"cert-manager-webhook-f4fb5df64-65hxg\" (UID: \"82008b0e-6099-474a-8526-f531424944ff\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.355408 4807 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hf9pp" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.364242 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:11 crc kubenswrapper[4807]: I1205 12:22:11.759501 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-65hxg"] Dec 05 12:22:11 crc kubenswrapper[4807]: W1205 12:22:11.767044 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82008b0e_6099_474a_8526_f531424944ff.slice/crio-8bffb17bd37072879501faeb3a32ad3cd6aaadf1ce9fcbbb024d8dd0c3686027 WatchSource:0}: Error finding container 8bffb17bd37072879501faeb3a32ad3cd6aaadf1ce9fcbbb024d8dd0c3686027: Status 404 returned error can't find the container with id 8bffb17bd37072879501faeb3a32ad3cd6aaadf1ce9fcbbb024d8dd0c3686027 Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.166852 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" event={"ID":"82008b0e-6099-474a-8526-f531424944ff","Type":"ContainerStarted","Data":"8bffb17bd37072879501faeb3a32ad3cd6aaadf1ce9fcbbb024d8dd0c3686027"} Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.386476 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k"] Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.387769 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.390191 4807 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-j7jrg" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.397560 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k"] Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.454050 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/574e8b05-7725-48f4-afdf-381468494129-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-tvt8k\" (UID: \"574e8b05-7725-48f4-afdf-381468494129\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.454095 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfqt9\" (UniqueName: \"kubernetes.io/projected/574e8b05-7725-48f4-afdf-381468494129-kube-api-access-kfqt9\") pod \"cert-manager-cainjector-855d9ccff4-tvt8k\" (UID: \"574e8b05-7725-48f4-afdf-381468494129\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.555690 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/574e8b05-7725-48f4-afdf-381468494129-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-tvt8k\" (UID: \"574e8b05-7725-48f4-afdf-381468494129\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.555743 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfqt9\" (UniqueName: \"kubernetes.io/projected/574e8b05-7725-48f4-afdf-381468494129-kube-api-access-kfqt9\") pod \"cert-manager-cainjector-855d9ccff4-tvt8k\" (UID: \"574e8b05-7725-48f4-afdf-381468494129\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.573486 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/574e8b05-7725-48f4-afdf-381468494129-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-tvt8k\" (UID: \"574e8b05-7725-48f4-afdf-381468494129\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.573641 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfqt9\" (UniqueName: \"kubernetes.io/projected/574e8b05-7725-48f4-afdf-381468494129-kube-api-access-kfqt9\") pod \"cert-manager-cainjector-855d9ccff4-tvt8k\" (UID: \"574e8b05-7725-48f4-afdf-381468494129\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:12 crc kubenswrapper[4807]: I1205 12:22:12.703486 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" Dec 05 12:22:13 crc kubenswrapper[4807]: I1205 12:22:13.007374 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k"] Dec 05 12:22:13 crc kubenswrapper[4807]: W1205 12:22:13.019346 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod574e8b05_7725_48f4_afdf_381468494129.slice/crio-9729ed4494d1961517786728025c52ae1804ccd3f66623b158df835f99ad1db9 WatchSource:0}: Error finding container 9729ed4494d1961517786728025c52ae1804ccd3f66623b158df835f99ad1db9: Status 404 returned error can't find the container with id 9729ed4494d1961517786728025c52ae1804ccd3f66623b158df835f99ad1db9 Dec 05 12:22:13 crc kubenswrapper[4807]: I1205 12:22:13.177517 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" event={"ID":"574e8b05-7725-48f4-afdf-381468494129","Type":"ContainerStarted","Data":"9729ed4494d1961517786728025c52ae1804ccd3f66623b158df835f99ad1db9"} Dec 05 12:22:21 crc kubenswrapper[4807]: I1205 12:22:21.249453 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" event={"ID":"82008b0e-6099-474a-8526-f531424944ff","Type":"ContainerStarted","Data":"5a1f1eb85b3f7be0558c65f57aa966c7ef638bdbc21ec3f65ce0854457967390"} Dec 05 12:22:21 crc kubenswrapper[4807]: I1205 12:22:21.250063 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:21 crc kubenswrapper[4807]: I1205 12:22:21.251552 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" event={"ID":"574e8b05-7725-48f4-afdf-381468494129","Type":"ContainerStarted","Data":"52a1cbb0b3758620e701c99f06315d928309fb7f6887175d98fd835d2b41765b"} Dec 05 12:22:21 crc kubenswrapper[4807]: I1205 12:22:21.292728 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-tvt8k" podStartSLOduration=1.952095927 podStartE2EDuration="9.292705949s" podCreationTimestamp="2025-12-05 12:22:12 +0000 UTC" firstStartedPulling="2025-12-05 12:22:13.02183421 +0000 UTC m=+962.515697479" lastFinishedPulling="2025-12-05 12:22:20.362444232 +0000 UTC m=+969.856307501" observedRunningTime="2025-12-05 12:22:21.289292814 +0000 UTC m=+970.783156103" watchObservedRunningTime="2025-12-05 12:22:21.292705949 +0000 UTC m=+970.786569228" Dec 05 12:22:21 crc kubenswrapper[4807]: I1205 12:22:21.316869 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" podStartSLOduration=1.738900009 podStartE2EDuration="10.316847114s" podCreationTimestamp="2025-12-05 12:22:11 +0000 UTC" firstStartedPulling="2025-12-05 12:22:11.769517687 +0000 UTC m=+961.263380946" lastFinishedPulling="2025-12-05 12:22:20.347464772 +0000 UTC m=+969.841328051" observedRunningTime="2025-12-05 12:22:21.311227336 +0000 UTC m=+970.805090615" watchObservedRunningTime="2025-12-05 12:22:21.316847114 +0000 UTC m=+970.810710383" Dec 05 12:22:22 crc kubenswrapper[4807]: I1205 12:22:22.466650 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:22:22 crc kubenswrapper[4807]: I1205 12:22:22.467007 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:22:26 crc kubenswrapper[4807]: I1205 12:22:26.367478 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-65hxg" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.246006 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-gh6vh"] Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.247278 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.249804 4807 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7khwz" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.262194 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-gh6vh"] Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.341224 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlxpj\" (UniqueName: \"kubernetes.io/projected/754c38f5-b65c-4ec6-905e-b5e1a3ca6cec-kube-api-access-jlxpj\") pod \"cert-manager-86cb77c54b-gh6vh\" (UID: \"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec\") " pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.341371 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/754c38f5-b65c-4ec6-905e-b5e1a3ca6cec-bound-sa-token\") pod \"cert-manager-86cb77c54b-gh6vh\" (UID: \"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec\") " pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.443206 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/754c38f5-b65c-4ec6-905e-b5e1a3ca6cec-bound-sa-token\") pod \"cert-manager-86cb77c54b-gh6vh\" (UID: \"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec\") " pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.443303 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlxpj\" (UniqueName: \"kubernetes.io/projected/754c38f5-b65c-4ec6-905e-b5e1a3ca6cec-kube-api-access-jlxpj\") pod \"cert-manager-86cb77c54b-gh6vh\" (UID: \"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec\") " pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.463662 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/754c38f5-b65c-4ec6-905e-b5e1a3ca6cec-bound-sa-token\") pod \"cert-manager-86cb77c54b-gh6vh\" (UID: \"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec\") " pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.464767 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlxpj\" (UniqueName: \"kubernetes.io/projected/754c38f5-b65c-4ec6-905e-b5e1a3ca6cec-kube-api-access-jlxpj\") pod \"cert-manager-86cb77c54b-gh6vh\" (UID: \"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec\") " pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.564631 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-gh6vh" Dec 05 12:22:29 crc kubenswrapper[4807]: I1205 12:22:29.966726 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-gh6vh"] Dec 05 12:22:30 crc kubenswrapper[4807]: I1205 12:22:30.305381 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-gh6vh" event={"ID":"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec","Type":"ContainerStarted","Data":"a731457205f69059be2df0f950dde055dc0b81567af892e9b12e87a14b1a88d4"} Dec 05 12:22:32 crc kubenswrapper[4807]: I1205 12:22:32.330484 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-gh6vh" event={"ID":"754c38f5-b65c-4ec6-905e-b5e1a3ca6cec","Type":"ContainerStarted","Data":"090160d5780f73627480406f21efdb747355d01d0d41571b4e2822c88533ade6"} Dec 05 12:22:32 crc kubenswrapper[4807]: I1205 12:22:32.343447 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-gh6vh" podStartSLOduration=3.3434291800000002 podStartE2EDuration="3.34342918s" podCreationTimestamp="2025-12-05 12:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:22:32.342564689 +0000 UTC m=+981.836427958" watchObservedRunningTime="2025-12-05 12:22:32.34342918 +0000 UTC m=+981.837292449" Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.850884 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r82zf"] Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.852421 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r82zf" Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.854969 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.855049 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.857649 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-qtg6h" Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.861107 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r82zf"] Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.880997 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xqdf\" (UniqueName: \"kubernetes.io/projected/f3fb2f22-4592-4559-a947-704fd50a10b9-kube-api-access-5xqdf\") pod \"openstack-operator-index-r82zf\" (UID: \"f3fb2f22-4592-4559-a947-704fd50a10b9\") " pod="openstack-operators/openstack-operator-index-r82zf" Dec 05 12:22:39 crc kubenswrapper[4807]: I1205 12:22:39.981923 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xqdf\" (UniqueName: \"kubernetes.io/projected/f3fb2f22-4592-4559-a947-704fd50a10b9-kube-api-access-5xqdf\") pod \"openstack-operator-index-r82zf\" (UID: \"f3fb2f22-4592-4559-a947-704fd50a10b9\") " pod="openstack-operators/openstack-operator-index-r82zf" Dec 05 12:22:40 crc kubenswrapper[4807]: I1205 12:22:40.009552 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xqdf\" (UniqueName: \"kubernetes.io/projected/f3fb2f22-4592-4559-a947-704fd50a10b9-kube-api-access-5xqdf\") pod \"openstack-operator-index-r82zf\" (UID: \"f3fb2f22-4592-4559-a947-704fd50a10b9\") " pod="openstack-operators/openstack-operator-index-r82zf" Dec 05 12:22:40 crc kubenswrapper[4807]: I1205 12:22:40.181087 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r82zf" Dec 05 12:22:40 crc kubenswrapper[4807]: I1205 12:22:40.390426 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r82zf"] Dec 05 12:22:41 crc kubenswrapper[4807]: I1205 12:22:41.384476 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r82zf" event={"ID":"f3fb2f22-4592-4559-a947-704fd50a10b9","Type":"ContainerStarted","Data":"17651adcace1fd96c29d407016658897da18b5f4eb54aff0f34d7bc813b97b37"} Dec 05 12:22:43 crc kubenswrapper[4807]: I1205 12:22:43.222873 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r82zf"] Dec 05 12:22:43 crc kubenswrapper[4807]: I1205 12:22:43.401264 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r82zf" event={"ID":"f3fb2f22-4592-4559-a947-704fd50a10b9","Type":"ContainerStarted","Data":"c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0"} Dec 05 12:22:43 crc kubenswrapper[4807]: I1205 12:22:43.419032 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r82zf" podStartSLOduration=2.241048697 podStartE2EDuration="4.419015814s" podCreationTimestamp="2025-12-05 12:22:39 +0000 UTC" firstStartedPulling="2025-12-05 12:22:40.397914491 +0000 UTC m=+989.891777760" lastFinishedPulling="2025-12-05 12:22:42.575881608 +0000 UTC m=+992.069744877" observedRunningTime="2025-12-05 12:22:43.415645169 +0000 UTC m=+992.909508438" watchObservedRunningTime="2025-12-05 12:22:43.419015814 +0000 UTC m=+992.912879083" Dec 05 12:22:43 crc kubenswrapper[4807]: I1205 12:22:43.830475 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-x89f7"] Dec 05 12:22:43 crc kubenswrapper[4807]: I1205 12:22:43.831634 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:43 crc kubenswrapper[4807]: I1205 12:22:43.839956 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-x89f7"] Dec 05 12:22:44 crc kubenswrapper[4807]: I1205 12:22:44.033785 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4z2s\" (UniqueName: \"kubernetes.io/projected/89e92216-a04d-416c-84b1-fbaa6e2e52b5-kube-api-access-w4z2s\") pod \"openstack-operator-index-x89f7\" (UID: \"89e92216-a04d-416c-84b1-fbaa6e2e52b5\") " pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:44 crc kubenswrapper[4807]: I1205 12:22:44.134589 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4z2s\" (UniqueName: \"kubernetes.io/projected/89e92216-a04d-416c-84b1-fbaa6e2e52b5-kube-api-access-w4z2s\") pod \"openstack-operator-index-x89f7\" (UID: \"89e92216-a04d-416c-84b1-fbaa6e2e52b5\") " pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:44 crc kubenswrapper[4807]: I1205 12:22:44.157382 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4z2s\" (UniqueName: \"kubernetes.io/projected/89e92216-a04d-416c-84b1-fbaa6e2e52b5-kube-api-access-w4z2s\") pod \"openstack-operator-index-x89f7\" (UID: \"89e92216-a04d-416c-84b1-fbaa6e2e52b5\") " pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:44 crc kubenswrapper[4807]: I1205 12:22:44.407455 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-r82zf" podUID="f3fb2f22-4592-4559-a947-704fd50a10b9" containerName="registry-server" containerID="cri-o://c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0" gracePeriod=2 Dec 05 12:22:44 crc kubenswrapper[4807]: I1205 12:22:44.450734 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:44 crc kubenswrapper[4807]: I1205 12:22:44.684717 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-x89f7"] Dec 05 12:22:44 crc kubenswrapper[4807]: W1205 12:22:44.723722 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89e92216_a04d_416c_84b1_fbaa6e2e52b5.slice/crio-c6b38c4d9f9b4f47500ca7fd6cc27bee977ecbb5d84db7c7103c309250745cb5 WatchSource:0}: Error finding container c6b38c4d9f9b4f47500ca7fd6cc27bee977ecbb5d84db7c7103c309250745cb5: Status 404 returned error can't find the container with id c6b38c4d9f9b4f47500ca7fd6cc27bee977ecbb5d84db7c7103c309250745cb5 Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.280695 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r82zf" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.417446 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-x89f7" event={"ID":"89e92216-a04d-416c-84b1-fbaa6e2e52b5","Type":"ContainerStarted","Data":"a625028994cafc4141639cf631ada588fec29fa84b9c2c2af1e07f1cadf08dd1"} Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.417501 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-x89f7" event={"ID":"89e92216-a04d-416c-84b1-fbaa6e2e52b5","Type":"ContainerStarted","Data":"c6b38c4d9f9b4f47500ca7fd6cc27bee977ecbb5d84db7c7103c309250745cb5"} Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.422466 4807 generic.go:334] "Generic (PLEG): container finished" podID="f3fb2f22-4592-4559-a947-704fd50a10b9" containerID="c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0" exitCode=0 Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.422507 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r82zf" event={"ID":"f3fb2f22-4592-4559-a947-704fd50a10b9","Type":"ContainerDied","Data":"c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0"} Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.422548 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r82zf" event={"ID":"f3fb2f22-4592-4559-a947-704fd50a10b9","Type":"ContainerDied","Data":"17651adcace1fd96c29d407016658897da18b5f4eb54aff0f34d7bc813b97b37"} Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.422564 4807 scope.go:117] "RemoveContainer" containerID="c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.422656 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r82zf" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.448769 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-x89f7" podStartSLOduration=2.220141161 podStartE2EDuration="2.448747757s" podCreationTimestamp="2025-12-05 12:22:43 +0000 UTC" firstStartedPulling="2025-12-05 12:22:44.728518427 +0000 UTC m=+994.222381696" lastFinishedPulling="2025-12-05 12:22:44.957125033 +0000 UTC m=+994.450988292" observedRunningTime="2025-12-05 12:22:45.432213279 +0000 UTC m=+994.926076548" watchObservedRunningTime="2025-12-05 12:22:45.448747757 +0000 UTC m=+994.942611076" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.449556 4807 scope.go:117] "RemoveContainer" containerID="c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.451317 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xqdf\" (UniqueName: \"kubernetes.io/projected/f3fb2f22-4592-4559-a947-704fd50a10b9-kube-api-access-5xqdf\") pod \"f3fb2f22-4592-4559-a947-704fd50a10b9\" (UID: \"f3fb2f22-4592-4559-a947-704fd50a10b9\") " Dec 05 12:22:45 crc kubenswrapper[4807]: E1205 12:22:45.453743 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0\": container with ID starting with c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0 not found: ID does not exist" containerID="c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.453820 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0"} err="failed to get container status \"c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0\": rpc error: code = NotFound desc = could not find container \"c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0\": container with ID starting with c0c6c92d8242a4f47d297f6fd2b69c8b0f6d7342b128bb59873d0175f51cc1a0 not found: ID does not exist" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.461460 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3fb2f22-4592-4559-a947-704fd50a10b9-kube-api-access-5xqdf" (OuterVolumeSpecName: "kube-api-access-5xqdf") pod "f3fb2f22-4592-4559-a947-704fd50a10b9" (UID: "f3fb2f22-4592-4559-a947-704fd50a10b9"). InnerVolumeSpecName "kube-api-access-5xqdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.553784 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xqdf\" (UniqueName: \"kubernetes.io/projected/f3fb2f22-4592-4559-a947-704fd50a10b9-kube-api-access-5xqdf\") on node \"crc\" DevicePath \"\"" Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.749357 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r82zf"] Dec 05 12:22:45 crc kubenswrapper[4807]: I1205 12:22:45.752735 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-r82zf"] Dec 05 12:22:47 crc kubenswrapper[4807]: I1205 12:22:47.242970 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3fb2f22-4592-4559-a947-704fd50a10b9" path="/var/lib/kubelet/pods/f3fb2f22-4592-4559-a947-704fd50a10b9/volumes" Dec 05 12:22:52 crc kubenswrapper[4807]: I1205 12:22:52.466108 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:22:52 crc kubenswrapper[4807]: I1205 12:22:52.466467 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:22:52 crc kubenswrapper[4807]: I1205 12:22:52.466515 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:22:52 crc kubenswrapper[4807]: I1205 12:22:52.467173 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9d6e7e624977c9e3e4da4ace9e23b2a19ff1d1ccbc78d45f4be87d54cc1e3e37"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:22:52 crc kubenswrapper[4807]: I1205 12:22:52.467231 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://9d6e7e624977c9e3e4da4ace9e23b2a19ff1d1ccbc78d45f4be87d54cc1e3e37" gracePeriod=600 Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.451496 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.452109 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.478161 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.480030 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="9d6e7e624977c9e3e4da4ace9e23b2a19ff1d1ccbc78d45f4be87d54cc1e3e37" exitCode=0 Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.480095 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"9d6e7e624977c9e3e4da4ace9e23b2a19ff1d1ccbc78d45f4be87d54cc1e3e37"} Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.480130 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"b47bace47d16c688e6065026bd16a58454816df8e9f3811211e0c6bea14f1f68"} Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.480148 4807 scope.go:117] "RemoveContainer" containerID="e2671acad109a1034ec97e5cc79cbdeae0c327f0a6312fb0cd334defd0fc9ebe" Dec 05 12:22:54 crc kubenswrapper[4807]: I1205 12:22:54.505637 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-x89f7" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.681417 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl"] Dec 05 12:23:00 crc kubenswrapper[4807]: E1205 12:23:00.682459 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fb2f22-4592-4559-a947-704fd50a10b9" containerName="registry-server" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.682480 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fb2f22-4592-4559-a947-704fd50a10b9" containerName="registry-server" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.682694 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fb2f22-4592-4559-a947-704fd50a10b9" containerName="registry-server" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.683769 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.688363 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-rsvqr" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.696200 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl"] Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.747722 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-util\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.747769 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp5x9\" (UniqueName: \"kubernetes.io/projected/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-kube-api-access-hp5x9\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.747865 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-bundle\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.848887 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-util\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.848936 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp5x9\" (UniqueName: \"kubernetes.io/projected/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-kube-api-access-hp5x9\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.848965 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-bundle\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.849362 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-bundle\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.849428 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-util\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:00 crc kubenswrapper[4807]: I1205 12:23:00.868086 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp5x9\" (UniqueName: \"kubernetes.io/projected/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-kube-api-access-hp5x9\") pod \"2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:01 crc kubenswrapper[4807]: I1205 12:23:01.057624 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:01 crc kubenswrapper[4807]: I1205 12:23:01.261022 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl"] Dec 05 12:23:01 crc kubenswrapper[4807]: I1205 12:23:01.532746 4807 generic.go:334] "Generic (PLEG): container finished" podID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerID="d17efbbcbc2bccc2a4e06e6479662fec1ecab824cafcc83f76cf7cbf6a921c79" exitCode=0 Dec 05 12:23:01 crc kubenswrapper[4807]: I1205 12:23:01.533014 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" event={"ID":"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6","Type":"ContainerDied","Data":"d17efbbcbc2bccc2a4e06e6479662fec1ecab824cafcc83f76cf7cbf6a921c79"} Dec 05 12:23:01 crc kubenswrapper[4807]: I1205 12:23:01.533048 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" event={"ID":"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6","Type":"ContainerStarted","Data":"6dc8d860075ad7ece138b1b2a7664a0303a27a05f6fa5cfffe8d64028059694a"} Dec 05 12:23:02 crc kubenswrapper[4807]: I1205 12:23:02.541211 4807 generic.go:334] "Generic (PLEG): container finished" podID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerID="8252308ee31728388580e3e0280a497bbc66b57fa41e1d35c56980bc07a3930d" exitCode=0 Dec 05 12:23:02 crc kubenswrapper[4807]: I1205 12:23:02.541266 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" event={"ID":"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6","Type":"ContainerDied","Data":"8252308ee31728388580e3e0280a497bbc66b57fa41e1d35c56980bc07a3930d"} Dec 05 12:23:03 crc kubenswrapper[4807]: I1205 12:23:03.550888 4807 generic.go:334] "Generic (PLEG): container finished" podID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerID="bf36d82549b1756f52ead141d0722760036c4e3d464cf1aa415a833ec965a370" exitCode=0 Dec 05 12:23:03 crc kubenswrapper[4807]: I1205 12:23:03.550932 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" event={"ID":"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6","Type":"ContainerDied","Data":"bf36d82549b1756f52ead141d0722760036c4e3d464cf1aa415a833ec965a370"} Dec 05 12:23:04 crc kubenswrapper[4807]: I1205 12:23:04.799835 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:04 crc kubenswrapper[4807]: I1205 12:23:04.899797 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-bundle\") pod \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " Dec 05 12:23:04 crc kubenswrapper[4807]: I1205 12:23:04.899906 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-util\") pod \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " Dec 05 12:23:04 crc kubenswrapper[4807]: I1205 12:23:04.899955 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp5x9\" (UniqueName: \"kubernetes.io/projected/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-kube-api-access-hp5x9\") pod \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\" (UID: \"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6\") " Dec 05 12:23:04 crc kubenswrapper[4807]: I1205 12:23:04.900849 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-bundle" (OuterVolumeSpecName: "bundle") pod "5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" (UID: "5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:23:04 crc kubenswrapper[4807]: I1205 12:23:04.907825 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-kube-api-access-hp5x9" (OuterVolumeSpecName: "kube-api-access-hp5x9") pod "5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" (UID: "5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6"). InnerVolumeSpecName "kube-api-access-hp5x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:23:04 crc kubenswrapper[4807]: I1205 12:23:04.915468 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-util" (OuterVolumeSpecName: "util") pod "5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" (UID: "5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:23:05 crc kubenswrapper[4807]: I1205 12:23:05.001546 4807 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-util\") on node \"crc\" DevicePath \"\"" Dec 05 12:23:05 crc kubenswrapper[4807]: I1205 12:23:05.001589 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp5x9\" (UniqueName: \"kubernetes.io/projected/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-kube-api-access-hp5x9\") on node \"crc\" DevicePath \"\"" Dec 05 12:23:05 crc kubenswrapper[4807]: I1205 12:23:05.001603 4807 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:23:05 crc kubenswrapper[4807]: I1205 12:23:05.567194 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" event={"ID":"5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6","Type":"ContainerDied","Data":"6dc8d860075ad7ece138b1b2a7664a0303a27a05f6fa5cfffe8d64028059694a"} Dec 05 12:23:05 crc kubenswrapper[4807]: I1205 12:23:05.567250 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dc8d860075ad7ece138b1b2a7664a0303a27a05f6fa5cfffe8d64028059694a" Dec 05 12:23:05 crc kubenswrapper[4807]: I1205 12:23:05.567351 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.340066 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7"] Dec 05 12:23:07 crc kubenswrapper[4807]: E1205 12:23:07.340840 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerName="util" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.340859 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerName="util" Dec 05 12:23:07 crc kubenswrapper[4807]: E1205 12:23:07.340887 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerName="extract" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.340896 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerName="extract" Dec 05 12:23:07 crc kubenswrapper[4807]: E1205 12:23:07.340931 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerName="pull" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.340939 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerName="pull" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.341425 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6" containerName="extract" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.342285 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.345549 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-rhwql" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.371459 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7"] Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.531428 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbwzf\" (UniqueName: \"kubernetes.io/projected/4e11a885-2cc7-43f4-9626-4aa22223946b-kube-api-access-bbwzf\") pod \"openstack-operator-controller-operator-574d9f8c97-6nmd7\" (UID: \"4e11a885-2cc7-43f4-9626-4aa22223946b\") " pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.632810 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbwzf\" (UniqueName: \"kubernetes.io/projected/4e11a885-2cc7-43f4-9626-4aa22223946b-kube-api-access-bbwzf\") pod \"openstack-operator-controller-operator-574d9f8c97-6nmd7\" (UID: \"4e11a885-2cc7-43f4-9626-4aa22223946b\") " pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.653116 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbwzf\" (UniqueName: \"kubernetes.io/projected/4e11a885-2cc7-43f4-9626-4aa22223946b-kube-api-access-bbwzf\") pod \"openstack-operator-controller-operator-574d9f8c97-6nmd7\" (UID: \"4e11a885-2cc7-43f4-9626-4aa22223946b\") " pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:23:07 crc kubenswrapper[4807]: I1205 12:23:07.665970 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:23:08 crc kubenswrapper[4807]: I1205 12:23:08.103884 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7"] Dec 05 12:23:08 crc kubenswrapper[4807]: W1205 12:23:08.104906 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e11a885_2cc7_43f4_9626_4aa22223946b.slice/crio-4ac787321eb8607c5ae411643484f10dfe7cb72ca1ec8ac29a1e8f9f0e41d2df WatchSource:0}: Error finding container 4ac787321eb8607c5ae411643484f10dfe7cb72ca1ec8ac29a1e8f9f0e41d2df: Status 404 returned error can't find the container with id 4ac787321eb8607c5ae411643484f10dfe7cb72ca1ec8ac29a1e8f9f0e41d2df Dec 05 12:23:08 crc kubenswrapper[4807]: I1205 12:23:08.587549 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" event={"ID":"4e11a885-2cc7-43f4-9626-4aa22223946b","Type":"ContainerStarted","Data":"4ac787321eb8607c5ae411643484f10dfe7cb72ca1ec8ac29a1e8f9f0e41d2df"} Dec 05 12:23:12 crc kubenswrapper[4807]: I1205 12:23:12.616988 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" event={"ID":"4e11a885-2cc7-43f4-9626-4aa22223946b","Type":"ContainerStarted","Data":"211fe2d123dcf47d2417ac38c8d2eff409d1f3ab7680aa964ade415089c5f56d"} Dec 05 12:23:12 crc kubenswrapper[4807]: I1205 12:23:12.617608 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:23:12 crc kubenswrapper[4807]: I1205 12:23:12.659729 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" podStartSLOduration=2.07255942 podStartE2EDuration="5.659708629s" podCreationTimestamp="2025-12-05 12:23:07 +0000 UTC" firstStartedPulling="2025-12-05 12:23:08.109098992 +0000 UTC m=+1017.602962271" lastFinishedPulling="2025-12-05 12:23:11.696248211 +0000 UTC m=+1021.190111480" observedRunningTime="2025-12-05 12:23:12.652377523 +0000 UTC m=+1022.146240862" watchObservedRunningTime="2025-12-05 12:23:12.659708629 +0000 UTC m=+1022.153571908" Dec 05 12:23:17 crc kubenswrapper[4807]: I1205 12:23:17.669931 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.004502 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.006000 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.008121 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-c4mc4" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.020197 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.021429 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.025254 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cnpfj" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.030259 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.035861 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.036749 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.037566 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjxks\" (UniqueName: \"kubernetes.io/projected/022ffaeb-bad6-48ee-be09-78fc4c515a99-kube-api-access-gjxks\") pod \"cinder-operator-controller-manager-859b6ccc6-ng4kv\" (UID: \"022ffaeb-bad6-48ee-be09-78fc4c515a99\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.037632 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmshf\" (UniqueName: \"kubernetes.io/projected/d2b94a74-f984-4360-8062-e42b82a7c401-kube-api-access-kmshf\") pod \"barbican-operator-controller-manager-7d9dfd778-zjzsd\" (UID: \"d2b94a74-f984-4360-8062-e42b82a7c401\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.038125 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-5bpc4" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.056033 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.061881 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.083590 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.084564 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.091641 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-769x6" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.104164 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.106181 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.108961 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-n9mnz" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.134592 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.140834 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjxks\" (UniqueName: \"kubernetes.io/projected/022ffaeb-bad6-48ee-be09-78fc4c515a99-kube-api-access-gjxks\") pod \"cinder-operator-controller-manager-859b6ccc6-ng4kv\" (UID: \"022ffaeb-bad6-48ee-be09-78fc4c515a99\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.140873 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr7m2\" (UniqueName: \"kubernetes.io/projected/d36f8a3a-450b-4824-826c-32283863f270-kube-api-access-dr7m2\") pod \"designate-operator-controller-manager-78b4bc895b-t82v9\" (UID: \"d36f8a3a-450b-4824-826c-32283863f270\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.140909 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmshf\" (UniqueName: \"kubernetes.io/projected/d2b94a74-f984-4360-8062-e42b82a7c401-kube-api-access-kmshf\") pod \"barbican-operator-controller-manager-7d9dfd778-zjzsd\" (UID: \"d2b94a74-f984-4360-8062-e42b82a7c401\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.140939 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fdsc\" (UniqueName: \"kubernetes.io/projected/ef9d3954-d526-4bd6-b806-e0ddb31994d3-kube-api-access-5fdsc\") pod \"heat-operator-controller-manager-5f64f6f8bb-j26nl\" (UID: \"ef9d3954-d526-4bd6-b806-e0ddb31994d3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.140968 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drntp\" (UniqueName: \"kubernetes.io/projected/c383d37a-acb0-463c-9980-1d2d46e06291-kube-api-access-drntp\") pod \"glance-operator-controller-manager-77987cd8cd-knvh4\" (UID: \"c383d37a-acb0-463c-9980-1d2d46e06291\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.146990 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.148426 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.152416 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8wgd9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.162588 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.187580 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.198756 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmshf\" (UniqueName: \"kubernetes.io/projected/d2b94a74-f984-4360-8062-e42b82a7c401-kube-api-access-kmshf\") pod \"barbican-operator-controller-manager-7d9dfd778-zjzsd\" (UID: \"d2b94a74-f984-4360-8062-e42b82a7c401\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.202181 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjxks\" (UniqueName: \"kubernetes.io/projected/022ffaeb-bad6-48ee-be09-78fc4c515a99-kube-api-access-gjxks\") pod \"cinder-operator-controller-manager-859b6ccc6-ng4kv\" (UID: \"022ffaeb-bad6-48ee-be09-78fc4c515a99\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.229916 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.231294 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.234711 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.234970 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4fb4t" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.278160 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr7m2\" (UniqueName: \"kubernetes.io/projected/d36f8a3a-450b-4824-826c-32283863f270-kube-api-access-dr7m2\") pod \"designate-operator-controller-manager-78b4bc895b-t82v9\" (UID: \"d36f8a3a-450b-4824-826c-32283863f270\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.279120 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llsnp\" (UniqueName: \"kubernetes.io/projected/7cd89edb-31e7-4c50-ba4d-c86f85fce855-kube-api-access-llsnp\") pod \"horizon-operator-controller-manager-68c6d99b8f-blrcs\" (UID: \"7cd89edb-31e7-4c50-ba4d-c86f85fce855\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.279400 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fdsc\" (UniqueName: \"kubernetes.io/projected/ef9d3954-d526-4bd6-b806-e0ddb31994d3-kube-api-access-5fdsc\") pod \"heat-operator-controller-manager-5f64f6f8bb-j26nl\" (UID: \"ef9d3954-d526-4bd6-b806-e0ddb31994d3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.280888 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz9gz\" (UniqueName: \"kubernetes.io/projected/4a8a470d-bdcf-4d64-b805-c996e88384ae-kube-api-access-pz9gz\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.281087 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drntp\" (UniqueName: \"kubernetes.io/projected/c383d37a-acb0-463c-9980-1d2d46e06291-kube-api-access-drntp\") pod \"glance-operator-controller-manager-77987cd8cd-knvh4\" (UID: \"c383d37a-acb0-463c-9980-1d2d46e06291\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.281349 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.344492 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr7m2\" (UniqueName: \"kubernetes.io/projected/d36f8a3a-450b-4824-826c-32283863f270-kube-api-access-dr7m2\") pod \"designate-operator-controller-manager-78b4bc895b-t82v9\" (UID: \"d36f8a3a-450b-4824-826c-32283863f270\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.345691 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.352071 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drntp\" (UniqueName: \"kubernetes.io/projected/c383d37a-acb0-463c-9980-1d2d46e06291-kube-api-access-drntp\") pod \"glance-operator-controller-manager-77987cd8cd-knvh4\" (UID: \"c383d37a-acb0-463c-9980-1d2d46e06291\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.352311 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.356713 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fdsc\" (UniqueName: \"kubernetes.io/projected/ef9d3954-d526-4bd6-b806-e0ddb31994d3-kube-api-access-5fdsc\") pod \"heat-operator-controller-manager-5f64f6f8bb-j26nl\" (UID: \"ef9d3954-d526-4bd6-b806-e0ddb31994d3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.359538 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.360565 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.362283 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-zfhl8" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.364930 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.370204 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.391020 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llsnp\" (UniqueName: \"kubernetes.io/projected/7cd89edb-31e7-4c50-ba4d-c86f85fce855-kube-api-access-llsnp\") pod \"horizon-operator-controller-manager-68c6d99b8f-blrcs\" (UID: \"7cd89edb-31e7-4c50-ba4d-c86f85fce855\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.391068 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz9gz\" (UniqueName: \"kubernetes.io/projected/4a8a470d-bdcf-4d64-b805-c996e88384ae-kube-api-access-pz9gz\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.391105 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.391129 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrscx\" (UniqueName: \"kubernetes.io/projected/5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86-kube-api-access-xrscx\") pod \"ironic-operator-controller-manager-6c548fd776-7pszd\" (UID: \"5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" Dec 05 12:23:37 crc kubenswrapper[4807]: E1205 12:23:37.391552 4807 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:37 crc kubenswrapper[4807]: E1205 12:23:37.391595 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert podName:4a8a470d-bdcf-4d64-b805-c996e88384ae nodeName:}" failed. No retries permitted until 2025-12-05 12:23:37.891580322 +0000 UTC m=+1047.385443591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert") pod "infra-operator-controller-manager-57548d458d-v6m8x" (UID: "4a8a470d-bdcf-4d64-b805-c996e88384ae") : secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.402593 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.419555 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.420378 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.420970 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llsnp\" (UniqueName: \"kubernetes.io/projected/7cd89edb-31e7-4c50-ba4d-c86f85fce855-kube-api-access-llsnp\") pod \"horizon-operator-controller-manager-68c6d99b8f-blrcs\" (UID: \"7cd89edb-31e7-4c50-ba4d-c86f85fce855\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.421494 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.426589 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2ptqs" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.432473 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.452936 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.454029 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.464807 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.468356 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-clgt2" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.471500 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.482462 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz9gz\" (UniqueName: \"kubernetes.io/projected/4a8a470d-bdcf-4d64-b805-c996e88384ae-kube-api-access-pz9gz\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.494005 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvgct\" (UniqueName: \"kubernetes.io/projected/65950bcd-6c98-49b5-b25f-f58cd0ee72bd-kube-api-access-zvgct\") pod \"manila-operator-controller-manager-7c79b5df47-5qfr7\" (UID: \"65950bcd-6c98-49b5-b25f-f58cd0ee72bd\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.494100 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsqnr\" (UniqueName: \"kubernetes.io/projected/64226f84-d145-459b-b326-acf7e87add59-kube-api-access-jsqnr\") pod \"keystone-operator-controller-manager-7765d96ddf-vl87h\" (UID: \"64226f84-d145-459b-b326-acf7e87add59\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.494175 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrscx\" (UniqueName: \"kubernetes.io/projected/5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86-kube-api-access-xrscx\") pod \"ironic-operator-controller-manager-6c548fd776-7pszd\" (UID: \"5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.499923 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.501120 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.508912 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-spkzc" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.529896 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrscx\" (UniqueName: \"kubernetes.io/projected/5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86-kube-api-access-xrscx\") pod \"ironic-operator-controller-manager-6c548fd776-7pszd\" (UID: \"5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.533067 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.567430 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.568909 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.572947 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-l66bm" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.578423 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.582081 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.583942 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-t86n9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.594132 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.597173 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkf68\" (UniqueName: \"kubernetes.io/projected/8817b3f7-5a96-42e2-9807-38a6742f5ac3-kube-api-access-zkf68\") pod \"mariadb-operator-controller-manager-56bbcc9d85-g5tvq\" (UID: \"8817b3f7-5a96-42e2-9807-38a6742f5ac3\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.597238 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mr74\" (UniqueName: \"kubernetes.io/projected/64763a6c-60c1-428e-baf9-040111b0057e-kube-api-access-6mr74\") pod \"nova-operator-controller-manager-697bc559fc-jzdm6\" (UID: \"64763a6c-60c1-428e-baf9-040111b0057e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.597302 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvgct\" (UniqueName: \"kubernetes.io/projected/65950bcd-6c98-49b5-b25f-f58cd0ee72bd-kube-api-access-zvgct\") pod \"manila-operator-controller-manager-7c79b5df47-5qfr7\" (UID: \"65950bcd-6c98-49b5-b25f-f58cd0ee72bd\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.597360 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkgwf\" (UniqueName: \"kubernetes.io/projected/568f4286-ea33-4e00-ab87-cdc930373223-kube-api-access-vkgwf\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rzhvm\" (UID: \"568f4286-ea33-4e00-ab87-cdc930373223\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.597403 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsqnr\" (UniqueName: \"kubernetes.io/projected/64226f84-d145-459b-b326-acf7e87add59-kube-api-access-jsqnr\") pod \"keystone-operator-controller-manager-7765d96ddf-vl87h\" (UID: \"64226f84-d145-459b-b326-acf7e87add59\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.631985 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvgct\" (UniqueName: \"kubernetes.io/projected/65950bcd-6c98-49b5-b25f-f58cd0ee72bd-kube-api-access-zvgct\") pod \"manila-operator-controller-manager-7c79b5df47-5qfr7\" (UID: \"65950bcd-6c98-49b5-b25f-f58cd0ee72bd\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.652940 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsqnr\" (UniqueName: \"kubernetes.io/projected/64226f84-d145-459b-b326-acf7e87add59-kube-api-access-jsqnr\") pod \"keystone-operator-controller-manager-7765d96ddf-vl87h\" (UID: \"64226f84-d145-459b-b326-acf7e87add59\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.684892 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.689003 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.698422 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkgwf\" (UniqueName: \"kubernetes.io/projected/568f4286-ea33-4e00-ab87-cdc930373223-kube-api-access-vkgwf\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rzhvm\" (UID: \"568f4286-ea33-4e00-ab87-cdc930373223\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.699819 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkf68\" (UniqueName: \"kubernetes.io/projected/8817b3f7-5a96-42e2-9807-38a6742f5ac3-kube-api-access-zkf68\") pod \"mariadb-operator-controller-manager-56bbcc9d85-g5tvq\" (UID: \"8817b3f7-5a96-42e2-9807-38a6742f5ac3\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.699883 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mr74\" (UniqueName: \"kubernetes.io/projected/64763a6c-60c1-428e-baf9-040111b0057e-kube-api-access-6mr74\") pod \"nova-operator-controller-manager-697bc559fc-jzdm6\" (UID: \"64763a6c-60c1-428e-baf9-040111b0057e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.707810 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.709173 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.718112 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-2thlv" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.721005 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.725267 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mr74\" (UniqueName: \"kubernetes.io/projected/64763a6c-60c1-428e-baf9-040111b0057e-kube-api-access-6mr74\") pod \"nova-operator-controller-manager-697bc559fc-jzdm6\" (UID: \"64763a6c-60c1-428e-baf9-040111b0057e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.731478 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkgwf\" (UniqueName: \"kubernetes.io/projected/568f4286-ea33-4e00-ab87-cdc930373223-kube-api-access-vkgwf\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rzhvm\" (UID: \"568f4286-ea33-4e00-ab87-cdc930373223\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.740391 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.741695 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.744269 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkf68\" (UniqueName: \"kubernetes.io/projected/8817b3f7-5a96-42e2-9807-38a6742f5ac3-kube-api-access-zkf68\") pod \"mariadb-operator-controller-manager-56bbcc9d85-g5tvq\" (UID: \"8817b3f7-5a96-42e2-9807-38a6742f5ac3\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.745783 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.745985 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xqxr8" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.756576 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.776233 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.796796 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.807811 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.807950 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r4ng\" (UniqueName: \"kubernetes.io/projected/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-kube-api-access-4r4ng\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.807978 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66lz2\" (UniqueName: \"kubernetes.io/projected/49f023c0-dfc7-4035-ae05-df3e6913e581-kube-api-access-66lz2\") pod \"octavia-operator-controller-manager-998648c74-t8bz9\" (UID: \"49f023c0-dfc7-4035-ae05-df3e6913e581\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.808141 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.809519 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.817910 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-9sl7j" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.821874 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.833983 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nclwx"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.835156 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.845607 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.855696 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nclwx"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.872751 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7blzz" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.873419 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.875353 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.885537 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.895614 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.897989 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.899705 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-kzmfb" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.904651 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh"] Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.910849 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-2hsqh" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915107 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkdv8\" (UniqueName: \"kubernetes.io/projected/63696f3a-63f2-475e-b5e1-9ed44bc6eb39-kube-api-access-fkdv8\") pod \"telemetry-operator-controller-manager-76cc84c6bb-mfzmp\" (UID: \"63696f3a-63f2-475e-b5e1-9ed44bc6eb39\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915179 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915248 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4bqq\" (UniqueName: \"kubernetes.io/projected/336d4305-fc09-4622-b6fd-ebf4fe266ad2-kube-api-access-f4bqq\") pod \"ovn-operator-controller-manager-b6456fdb6-99xgh\" (UID: \"336d4305-fc09-4622-b6fd-ebf4fe266ad2\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915280 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r4ng\" (UniqueName: \"kubernetes.io/projected/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-kube-api-access-4r4ng\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915320 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww49j\" (UniqueName: \"kubernetes.io/projected/85952cd3-56c3-49d2-9f15-cd1f872a6c7c-kube-api-access-ww49j\") pod \"swift-operator-controller-manager-5f8c65bbfc-zbdbh\" (UID: \"85952cd3-56c3-49d2-9f15-cd1f872a6c7c\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915340 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66lz2\" (UniqueName: \"kubernetes.io/projected/49f023c0-dfc7-4035-ae05-df3e6913e581-kube-api-access-66lz2\") pod \"octavia-operator-controller-manager-998648c74-t8bz9\" (UID: \"49f023c0-dfc7-4035-ae05-df3e6913e581\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915402 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.915435 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzhq\" (UniqueName: \"kubernetes.io/projected/2f6ae961-a4b7-4d96-8669-2b0c1653ddc3-kube-api-access-2nzhq\") pod \"placement-operator-controller-manager-78f8948974-nclwx\" (UID: \"2f6ae961-a4b7-4d96-8669-2b0c1653ddc3\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" Dec 05 12:23:37 crc kubenswrapper[4807]: E1205 12:23:37.922206 4807 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:37 crc kubenswrapper[4807]: E1205 12:23:37.922318 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert podName:f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a nodeName:}" failed. No retries permitted until 2025-12-05 12:23:38.422293833 +0000 UTC m=+1047.916157102 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" (UID: "f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:37 crc kubenswrapper[4807]: E1205 12:23:37.925374 4807 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:37 crc kubenswrapper[4807]: E1205 12:23:37.925600 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert podName:4a8a470d-bdcf-4d64-b805-c996e88384ae nodeName:}" failed. No retries permitted until 2025-12-05 12:23:38.925417333 +0000 UTC m=+1048.419280602 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert") pod "infra-operator-controller-manager-57548d458d-v6m8x" (UID: "4a8a470d-bdcf-4d64-b805-c996e88384ae") : secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.933616 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.934457 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" Dec 05 12:23:37 crc kubenswrapper[4807]: I1205 12:23:37.963070 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.022148 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkdv8\" (UniqueName: \"kubernetes.io/projected/63696f3a-63f2-475e-b5e1-9ed44bc6eb39-kube-api-access-fkdv8\") pod \"telemetry-operator-controller-manager-76cc84c6bb-mfzmp\" (UID: \"63696f3a-63f2-475e-b5e1-9ed44bc6eb39\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.026373 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4bqq\" (UniqueName: \"kubernetes.io/projected/336d4305-fc09-4622-b6fd-ebf4fe266ad2-kube-api-access-f4bqq\") pod \"ovn-operator-controller-manager-b6456fdb6-99xgh\" (UID: \"336d4305-fc09-4622-b6fd-ebf4fe266ad2\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.026443 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww49j\" (UniqueName: \"kubernetes.io/projected/85952cd3-56c3-49d2-9f15-cd1f872a6c7c-kube-api-access-ww49j\") pod \"swift-operator-controller-manager-5f8c65bbfc-zbdbh\" (UID: \"85952cd3-56c3-49d2-9f15-cd1f872a6c7c\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.026546 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzhq\" (UniqueName: \"kubernetes.io/projected/2f6ae961-a4b7-4d96-8669-2b0c1653ddc3-kube-api-access-2nzhq\") pod \"placement-operator-controller-manager-78f8948974-nclwx\" (UID: \"2f6ae961-a4b7-4d96-8669-2b0c1653ddc3\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.031789 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66lz2\" (UniqueName: \"kubernetes.io/projected/49f023c0-dfc7-4035-ae05-df3e6913e581-kube-api-access-66lz2\") pod \"octavia-operator-controller-manager-998648c74-t8bz9\" (UID: \"49f023c0-dfc7-4035-ae05-df3e6913e581\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.032787 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r4ng\" (UniqueName: \"kubernetes.io/projected/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-kube-api-access-4r4ng\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.048673 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.082419 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.095575 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.095671 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.100358 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-qhb47" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.104440 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzhq\" (UniqueName: \"kubernetes.io/projected/2f6ae961-a4b7-4d96-8669-2b0c1653ddc3-kube-api-access-2nzhq\") pod \"placement-operator-controller-manager-78f8948974-nclwx\" (UID: \"2f6ae961-a4b7-4d96-8669-2b0c1653ddc3\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.107738 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkdv8\" (UniqueName: \"kubernetes.io/projected/63696f3a-63f2-475e-b5e1-9ed44bc6eb39-kube-api-access-fkdv8\") pod \"telemetry-operator-controller-manager-76cc84c6bb-mfzmp\" (UID: \"63696f3a-63f2-475e-b5e1-9ed44bc6eb39\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.126307 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww49j\" (UniqueName: \"kubernetes.io/projected/85952cd3-56c3-49d2-9f15-cd1f872a6c7c-kube-api-access-ww49j\") pod \"swift-operator-controller-manager-5f8c65bbfc-zbdbh\" (UID: \"85952cd3-56c3-49d2-9f15-cd1f872a6c7c\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.126887 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4bqq\" (UniqueName: \"kubernetes.io/projected/336d4305-fc09-4622-b6fd-ebf4fe266ad2-kube-api-access-f4bqq\") pod \"ovn-operator-controller-manager-b6456fdb6-99xgh\" (UID: \"336d4305-fc09-4622-b6fd-ebf4fe266ad2\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.127754 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-295f4\" (UniqueName: \"kubernetes.io/projected/5be4b980-a6ab-4499-87ec-0aff47a7a917-kube-api-access-295f4\") pod \"test-operator-controller-manager-5854674fcc-g2mx8\" (UID: \"5be4b980-a6ab-4499-87ec-0aff47a7a917\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.145290 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.166591 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.167966 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.175039 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.180932 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-x59cb" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.187668 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.229723 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbgcp\" (UniqueName: \"kubernetes.io/projected/3e3e2978-5f36-49af-ac8f-3ca92ada318e-kube-api-access-dbgcp\") pod \"watcher-operator-controller-manager-6c9545865c-bdpck\" (UID: \"3e3e2978-5f36-49af-ac8f-3ca92ada318e\") " pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.229793 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-295f4\" (UniqueName: \"kubernetes.io/projected/5be4b980-a6ab-4499-87ec-0aff47a7a917-kube-api-access-295f4\") pod \"test-operator-controller-manager-5854674fcc-g2mx8\" (UID: \"5be4b980-a6ab-4499-87ec-0aff47a7a917\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.234774 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.236185 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.246460 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.247057 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-llgj7" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.247369 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.263336 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.289508 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-295f4\" (UniqueName: \"kubernetes.io/projected/5be4b980-a6ab-4499-87ec-0aff47a7a917-kube-api-access-295f4\") pod \"test-operator-controller-manager-5854674fcc-g2mx8\" (UID: \"5be4b980-a6ab-4499-87ec-0aff47a7a917\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.301987 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.310358 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.328472 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.329251 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-7lbxd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.346327 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.346369 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbgcp\" (UniqueName: \"kubernetes.io/projected/3e3e2978-5f36-49af-ac8f-3ca92ada318e-kube-api-access-dbgcp\") pod \"watcher-operator-controller-manager-6c9545865c-bdpck\" (UID: \"3e3e2978-5f36-49af-ac8f-3ca92ada318e\") " pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.346414 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh47f\" (UniqueName: \"kubernetes.io/projected/bc48b693-8a75-4dea-a3cf-3dd053a8a500-kube-api-access-gh47f\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.349682 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.350885 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t7h8\" (UniqueName: \"kubernetes.io/projected/87b435a4-0ca7-4a37-9736-40ff511bfb43-kube-api-access-6t7h8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-r2q7v\" (UID: \"87b435a4-0ca7-4a37-9736-40ff511bfb43\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.350344 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.383770 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbgcp\" (UniqueName: \"kubernetes.io/projected/3e3e2978-5f36-49af-ac8f-3ca92ada318e-kube-api-access-dbgcp\") pod \"watcher-operator-controller-manager-6c9545865c-bdpck\" (UID: \"3e3e2978-5f36-49af-ac8f-3ca92ada318e\") " pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.453516 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.453580 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.453610 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh47f\" (UniqueName: \"kubernetes.io/projected/bc48b693-8a75-4dea-a3cf-3dd053a8a500-kube-api-access-gh47f\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.453633 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.453694 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t7h8\" (UniqueName: \"kubernetes.io/projected/87b435a4-0ca7-4a37-9736-40ff511bfb43-kube-api-access-6t7h8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-r2q7v\" (UID: \"87b435a4-0ca7-4a37-9736-40ff511bfb43\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.454057 4807 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.454094 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:38.954080442 +0000 UTC m=+1048.447943711 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.454216 4807 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.454236 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert podName:f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a nodeName:}" failed. No retries permitted until 2025-12-05 12:23:39.454229326 +0000 UTC m=+1048.948092595 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" (UID: "f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.454748 4807 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.454809 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:38.954800061 +0000 UTC m=+1048.448663330 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "metrics-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.465804 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.466547 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.479519 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.480928 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh47f\" (UniqueName: \"kubernetes.io/projected/bc48b693-8a75-4dea-a3cf-3dd053a8a500-kube-api-access-gh47f\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.507730 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t7h8\" (UniqueName: \"kubernetes.io/projected/87b435a4-0ca7-4a37-9736-40ff511bfb43-kube-api-access-6t7h8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-r2q7v\" (UID: \"87b435a4-0ca7-4a37-9736-40ff511bfb43\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.534409 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.837360 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.892139 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv"] Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.977197 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.977264 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:38 crc kubenswrapper[4807]: I1205 12:23:38.977301 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.977440 4807 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.977491 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert podName:4a8a470d-bdcf-4d64-b805-c996e88384ae nodeName:}" failed. No retries permitted until 2025-12-05 12:23:40.977475759 +0000 UTC m=+1050.471339028 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert") pod "infra-operator-controller-manager-57548d458d-v6m8x" (UID: "4a8a470d-bdcf-4d64-b805-c996e88384ae") : secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.977870 4807 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.977901 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:39.977892389 +0000 UTC m=+1049.471755658 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "webhook-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.977946 4807 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 12:23:38 crc kubenswrapper[4807]: E1205 12:23:38.977964 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:39.97795836 +0000 UTC m=+1049.471821629 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "metrics-server-cert" not found Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.132715 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.146419 4807 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.149280 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.169379 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs"] Dec 05 12:23:39 crc kubenswrapper[4807]: W1205 12:23:39.174097 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cd89edb_31e7_4c50_ba4d_c86f85fce855.slice/crio-2d87120ba39490d663cfb8787a20c995477218b8e82a82cbd536a946fd45e6fc WatchSource:0}: Error finding container 2d87120ba39490d663cfb8787a20c995477218b8e82a82cbd536a946fd45e6fc: Status 404 returned error can't find the container with id 2d87120ba39490d663cfb8787a20c995477218b8e82a82cbd536a946fd45e6fc Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.504775 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:39 crc kubenswrapper[4807]: E1205 12:23:39.504940 4807 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:39 crc kubenswrapper[4807]: E1205 12:23:39.505218 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert podName:f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a nodeName:}" failed. No retries permitted until 2025-12-05 12:23:41.505200924 +0000 UTC m=+1050.999064193 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" (UID: "f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.656776 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.676097 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nclwx"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.703433 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.723211 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.784151 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.810439 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.825944 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.864601 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" event={"ID":"64763a6c-60c1-428e-baf9-040111b0057e","Type":"ContainerStarted","Data":"6d8666c29f70b7b4d3834a75e9db3898e20f5f1993eb269d05bfa602cf9c01db"} Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.887239 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" event={"ID":"7cd89edb-31e7-4c50-ba4d-c86f85fce855","Type":"ContainerStarted","Data":"2d87120ba39490d663cfb8787a20c995477218b8e82a82cbd536a946fd45e6fc"} Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.909737 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" event={"ID":"022ffaeb-bad6-48ee-be09-78fc4c515a99","Type":"ContainerStarted","Data":"1f32149f85dcfb8cc3bfd399de7e30c323e4fecb714e0d04cb02da5347a47f33"} Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.922578 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.939701 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.946012 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.948786 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" event={"ID":"d2b94a74-f984-4360-8062-e42b82a7c401","Type":"ContainerStarted","Data":"5dd9f2473ee9a967c26206f76c2f9e1cb5f4a3473640998aa0bd3d968baa44b9"} Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.958766 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" event={"ID":"2f6ae961-a4b7-4d96-8669-2b0c1653ddc3","Type":"ContainerStarted","Data":"0e8e4daadad844f56b79ad3287afd9b372298e10b36f033842f581206cf6ac6e"} Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.987456 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9"] Dec 05 12:23:39 crc kubenswrapper[4807]: W1205 12:23:39.989647 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65950bcd_6c98_49b5_b25f_f58cd0ee72bd.slice/crio-d83b15372f6d0c7c5d0bfca2f5420df4aff921b6030501934b402f998049047b WatchSource:0}: Error finding container d83b15372f6d0c7c5d0bfca2f5420df4aff921b6030501934b402f998049047b: Status 404 returned error can't find the container with id d83b15372f6d0c7c5d0bfca2f5420df4aff921b6030501934b402f998049047b Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.995788 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh"] Dec 05 12:23:39 crc kubenswrapper[4807]: I1205 12:23:39.999488 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" event={"ID":"d36f8a3a-450b-4824-826c-32283863f270","Type":"ContainerStarted","Data":"bbed6982d2e1738db665f7a0025834ac2baf75f4c5795740022fa3f2f7294031"} Dec 05 12:23:40 crc kubenswrapper[4807]: I1205 12:23:40.007811 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v"] Dec 05 12:23:40 crc kubenswrapper[4807]: I1205 12:23:40.010207 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp"] Dec 05 12:23:40 crc kubenswrapper[4807]: I1205 12:23:40.021842 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" event={"ID":"64226f84-d145-459b-b326-acf7e87add59","Type":"ContainerStarted","Data":"d0121fdfeb0f8962cd3ee899786a9a9bcbcaac25d6bb5eb819876a393daac029"} Dec 05 12:23:40 crc kubenswrapper[4807]: I1205 12:23:40.022302 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:40 crc kubenswrapper[4807]: I1205 12:23:40.022356 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.022433 4807 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.022503 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:42.022483136 +0000 UTC m=+1051.516346405 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "webhook-server-cert" not found Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.022510 4807 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.022611 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:42.0225962 +0000 UTC m=+1051.516459469 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "metrics-server-cert" not found Dec 05 12:23:40 crc kubenswrapper[4807]: W1205 12:23:40.023903 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85952cd3_56c3_49d2_9f15_cd1f872a6c7c.slice/crio-5ca906e8aec4e8f7c30f905057bd484046868bfd68969a6aebd62a931718965e WatchSource:0}: Error finding container 5ca906e8aec4e8f7c30f905057bd484046868bfd68969a6aebd62a931718965e: Status 404 returned error can't find the container with id 5ca906e8aec4e8f7c30f905057bd484046868bfd68969a6aebd62a931718965e Dec 05 12:23:40 crc kubenswrapper[4807]: W1205 12:23:40.026689 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87b435a4_0ca7_4a37_9736_40ff511bfb43.slice/crio-dd933266435c895143898ff01127179d235042a5e99aa383e4668bc2ec7c6f56 WatchSource:0}: Error finding container dd933266435c895143898ff01127179d235042a5e99aa383e4668bc2ec7c6f56: Status 404 returned error can't find the container with id dd933266435c895143898ff01127179d235042a5e99aa383e4668bc2ec7c6f56 Dec 05 12:23:40 crc kubenswrapper[4807]: I1205 12:23:40.029674 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck"] Dec 05 12:23:40 crc kubenswrapper[4807]: I1205 12:23:40.032882 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" event={"ID":"c383d37a-acb0-463c-9980-1d2d46e06291","Type":"ContainerStarted","Data":"89cd63ac2edf960785c9114992361fbcc05877ec958ee86ac5bb7ed080b51740"} Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.042889 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ww49j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-zbdbh_openstack-operators(85952cd3-56c3-49d2-9f15-cd1f872a6c7c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.043004 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fkdv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-mfzmp_openstack-operators(63696f3a-63f2-475e-b5e1-9ed44bc6eb39): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.043600 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6t7h8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-r2q7v_openstack-operators(87b435a4-0ca7-4a37-9736-40ff511bfb43): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.045071 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fkdv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-mfzmp_openstack-operators(63696f3a-63f2-475e-b5e1-9ed44bc6eb39): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.045370 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ww49j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-zbdbh_openstack-operators(85952cd3-56c3-49d2-9f15-cd1f872a6c7c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.045721 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" podUID="87b435a4-0ca7-4a37-9736-40ff511bfb43" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.046598 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" podUID="85952cd3-56c3-49d2-9f15-cd1f872a6c7c" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.046658 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" podUID="63696f3a-63f2-475e-b5e1-9ed44bc6eb39" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.056355 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66lz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-t8bz9_openstack-operators(49f023c0-dfc7-4035-ae05-df3e6913e581): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.063964 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.20:5001/openstack-k8s-operators/watcher-operator:d23b8876e1bcf18983498fca8ec9314bc8124a8c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dbgcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c9545865c-bdpck_openstack-operators(3e3e2978-5f36-49af-ac8f-3ca92ada318e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.064005 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66lz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-t8bz9_openstack-operators(49f023c0-dfc7-4035-ae05-df3e6913e581): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.065758 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" podUID="49f023c0-dfc7-4035-ae05-df3e6913e581" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.066708 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dbgcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c9545865c-bdpck_openstack-operators(3e3e2978-5f36-49af-ac8f-3ca92ada318e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:40 crc kubenswrapper[4807]: E1205 12:23:40.068666 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.036696 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.036881 4807 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.037175 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert podName:4a8a470d-bdcf-4d64-b805-c996e88384ae nodeName:}" failed. No retries permitted until 2025-12-05 12:23:45.037157047 +0000 UTC m=+1054.531020316 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert") pod "infra-operator-controller-manager-57548d458d-v6m8x" (UID: "4a8a470d-bdcf-4d64-b805-c996e88384ae") : secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.059314 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" event={"ID":"ef9d3954-d526-4bd6-b806-e0ddb31994d3","Type":"ContainerStarted","Data":"fa627eae77efcd49e6bd277f5997e681df7b533a36c947e84138bda99b65767e"} Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.062680 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" event={"ID":"85952cd3-56c3-49d2-9f15-cd1f872a6c7c","Type":"ContainerStarted","Data":"5ca906e8aec4e8f7c30f905057bd484046868bfd68969a6aebd62a931718965e"} Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.065056 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" podUID="85952cd3-56c3-49d2-9f15-cd1f872a6c7c" Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.073426 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" event={"ID":"65950bcd-6c98-49b5-b25f-f58cd0ee72bd","Type":"ContainerStarted","Data":"d83b15372f6d0c7c5d0bfca2f5420df4aff921b6030501934b402f998049047b"} Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.078982 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" event={"ID":"336d4305-fc09-4622-b6fd-ebf4fe266ad2","Type":"ContainerStarted","Data":"d9a4799d06c1887ce5e926e35b92b196d40c0500f15a312936db929e5e4e263b"} Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.080634 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" event={"ID":"63696f3a-63f2-475e-b5e1-9ed44bc6eb39","Type":"ContainerStarted","Data":"5b01b0eb557ce06276fc3f7fe657ffd28d8ca11de7783fae4a74db1a31a4b39b"} Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.085561 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" podUID="63696f3a-63f2-475e-b5e1-9ed44bc6eb39" Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.086567 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" event={"ID":"5be4b980-a6ab-4499-87ec-0aff47a7a917","Type":"ContainerStarted","Data":"28418a0c03fe4dcd1b5ffe0ca4326f14b75dfcaf05cfcf6c161911103a67adb9"} Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.092323 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" event={"ID":"8817b3f7-5a96-42e2-9807-38a6742f5ac3","Type":"ContainerStarted","Data":"7019d5456fdf75a9601ccaec0f759548e74cdf9afd90865c014f901413c11b73"} Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.107769 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" event={"ID":"3e3e2978-5f36-49af-ac8f-3ca92ada318e","Type":"ContainerStarted","Data":"49e00b4a435eb31a6b292328884746d2ca775cab1b3862edd0c1f06476b281b8"} Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.125698 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.20:5001/openstack-k8s-operators/watcher-operator:d23b8876e1bcf18983498fca8ec9314bc8124a8c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.131807 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" event={"ID":"568f4286-ea33-4e00-ab87-cdc930373223","Type":"ContainerStarted","Data":"4675511e1fb34fab51b5e8466eeaedd281661dc90f4453c294db5462ede7d0bf"} Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.136702 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" event={"ID":"87b435a4-0ca7-4a37-9736-40ff511bfb43","Type":"ContainerStarted","Data":"dd933266435c895143898ff01127179d235042a5e99aa383e4668bc2ec7c6f56"} Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.141894 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" podUID="87b435a4-0ca7-4a37-9736-40ff511bfb43" Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.146822 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" event={"ID":"5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86","Type":"ContainerStarted","Data":"71ceba35e301cd5e55c7111e6a300f12c1252c732b53616dfe81af39e6a4a674"} Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.162012 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" event={"ID":"49f023c0-dfc7-4035-ae05-df3e6913e581","Type":"ContainerStarted","Data":"ef6ce1bae9f7a97569f15729926ff4a16262c7ba4f4d317cd3cf375c81b0ce51"} Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.174827 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" podUID="49f023c0-dfc7-4035-ae05-df3e6913e581" Dec 05 12:23:41 crc kubenswrapper[4807]: I1205 12:23:41.560792 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.560961 4807 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:41 crc kubenswrapper[4807]: E1205 12:23:41.561018 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert podName:f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a nodeName:}" failed. No retries permitted until 2025-12-05 12:23:45.560990615 +0000 UTC m=+1055.054853884 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" (UID: "f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:42 crc kubenswrapper[4807]: I1205 12:23:42.067909 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:42 crc kubenswrapper[4807]: I1205 12:23:42.067996 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.068141 4807 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.068149 4807 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.068208 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:46.068190923 +0000 UTC m=+1055.562054192 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "metrics-server-cert" not found Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.068237 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:46.068215993 +0000 UTC m=+1055.562079312 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "webhook-server-cert" not found Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.175185 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" podUID="87b435a4-0ca7-4a37-9736-40ff511bfb43" Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.175698 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" podUID="63696f3a-63f2-475e-b5e1-9ed44bc6eb39" Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.176049 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.20:5001/openstack-k8s-operators/watcher-operator:d23b8876e1bcf18983498fca8ec9314bc8124a8c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.176212 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" podUID="85952cd3-56c3-49d2-9f15-cd1f872a6c7c" Dec 05 12:23:42 crc kubenswrapper[4807]: E1205 12:23:42.176964 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" podUID="49f023c0-dfc7-4035-ae05-df3e6913e581" Dec 05 12:23:45 crc kubenswrapper[4807]: I1205 12:23:45.134967 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:45 crc kubenswrapper[4807]: E1205 12:23:45.135392 4807 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:45 crc kubenswrapper[4807]: E1205 12:23:45.135439 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert podName:4a8a470d-bdcf-4d64-b805-c996e88384ae nodeName:}" failed. No retries permitted until 2025-12-05 12:23:53.135423942 +0000 UTC m=+1062.629287201 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert") pod "infra-operator-controller-manager-57548d458d-v6m8x" (UID: "4a8a470d-bdcf-4d64-b805-c996e88384ae") : secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:45 crc kubenswrapper[4807]: I1205 12:23:45.645005 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:45 crc kubenswrapper[4807]: E1205 12:23:45.645178 4807 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:45 crc kubenswrapper[4807]: E1205 12:23:45.645242 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert podName:f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a nodeName:}" failed. No retries permitted until 2025-12-05 12:23:53.645222735 +0000 UTC m=+1063.139086014 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" (UID: "f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:46 crc kubenswrapper[4807]: I1205 12:23:46.153214 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:46 crc kubenswrapper[4807]: I1205 12:23:46.153351 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:46 crc kubenswrapper[4807]: E1205 12:23:46.153384 4807 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 12:23:46 crc kubenswrapper[4807]: E1205 12:23:46.153452 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:54.153432858 +0000 UTC m=+1063.647296127 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "webhook-server-cert" not found Dec 05 12:23:46 crc kubenswrapper[4807]: E1205 12:23:46.153519 4807 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 12:23:46 crc kubenswrapper[4807]: E1205 12:23:46.153629 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:23:54.153605682 +0000 UTC m=+1063.647469011 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "metrics-server-cert" not found Dec 05 12:23:53 crc kubenswrapper[4807]: I1205 12:23:53.167346 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.167502 4807 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.167905 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert podName:4a8a470d-bdcf-4d64-b805-c996e88384ae nodeName:}" failed. No retries permitted until 2025-12-05 12:24:09.167888116 +0000 UTC m=+1078.661751385 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert") pod "infra-operator-controller-manager-57548d458d-v6m8x" (UID: "4a8a470d-bdcf-4d64-b805-c996e88384ae") : secret "infra-operator-webhook-server-cert" not found Dec 05 12:23:53 crc kubenswrapper[4807]: I1205 12:23:53.678760 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.679257 4807 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.679309 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert podName:f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a nodeName:}" failed. No retries permitted until 2025-12-05 12:24:09.679293969 +0000 UTC m=+1079.173157238 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" (UID: "f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.703056 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zkf68,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-g5tvq_openstack-operators(8817b3f7-5a96-42e2-9807-38a6742f5ac3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.705210 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" podUID="8817b3f7-5a96-42e2-9807-38a6742f5ac3" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.723722 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dr7m2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-t82v9_openstack-operators(d36f8a3a-450b-4824-826c-32283863f270): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.724801 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vkgwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-rzhvm_openstack-operators(568f4286-ea33-4e00-ab87-cdc930373223): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.725797 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" podUID="d36f8a3a-450b-4824-826c-32283863f270" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.725935 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" podUID="568f4286-ea33-4e00-ab87-cdc930373223" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.742120 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmshf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-zjzsd_openstack-operators(d2b94a74-f984-4360-8062-e42b82a7c401): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 12:23:53 crc kubenswrapper[4807]: E1205 12:23:53.743594 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" podUID="d2b94a74-f984-4360-8062-e42b82a7c401" Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.187562 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.187680 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.187823 4807 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.187871 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:24:10.187856091 +0000 UTC m=+1079.681719350 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "webhook-server-cert" not found Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.188182 4807 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.188208 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs podName:bc48b693-8a75-4dea-a3cf-3dd053a8a500 nodeName:}" failed. No retries permitted until 2025-12-05 12:24:10.18820022 +0000 UTC m=+1079.682063489 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs") pod "openstack-operator-controller-manager-fcb8bd8db-sbftd" (UID: "bc48b693-8a75-4dea-a3cf-3dd053a8a500") : secret "metrics-server-cert" not found Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.286054 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" event={"ID":"022ffaeb-bad6-48ee-be09-78fc4c515a99","Type":"ContainerStarted","Data":"ce6da2de36a7c4cbe3a6789bb6767450c6f7aaf20dcb846e14016cde55602d0a"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.304307 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" event={"ID":"d2b94a74-f984-4360-8062-e42b82a7c401","Type":"ContainerStarted","Data":"f1ba27298ace3cd120a63babb73d810c8fbc57b0a2020e7e98233fff9a57624e"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.304387 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.305892 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" podUID="d2b94a74-f984-4360-8062-e42b82a7c401" Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.317854 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" event={"ID":"5be4b980-a6ab-4499-87ec-0aff47a7a917","Type":"ContainerStarted","Data":"dc04fb28581de0992c888f045fd5ae3ba2bc4e63338b930e8113126826c3fb62"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.336811 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" event={"ID":"65950bcd-6c98-49b5-b25f-f58cd0ee72bd","Type":"ContainerStarted","Data":"e53e48a1868383f929647ef7ded02fdd495227ae558a68dcbdd1409cf2271a40"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.358342 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" event={"ID":"5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86","Type":"ContainerStarted","Data":"532f9401cb6921c02312a6906005586895b1c905528c91f217e05958098de2aa"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.392224 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" event={"ID":"336d4305-fc09-4622-b6fd-ebf4fe266ad2","Type":"ContainerStarted","Data":"429abff1e5c0de967839cc8d646ba582e3369f448746276a1eead67d61f697c3"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.408347 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" event={"ID":"64763a6c-60c1-428e-baf9-040111b0057e","Type":"ContainerStarted","Data":"760cf3c3e6bd380c4c7898053e68e4c0991b8c809aa6279c282be03388ed7863"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.432778 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" event={"ID":"8817b3f7-5a96-42e2-9807-38a6742f5ac3","Type":"ContainerStarted","Data":"72f66489c5102b8ba6d37e22f39a65b2ffcc2e213864ff87e92c3bf5065749fc"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.433648 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.437979 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" podUID="8817b3f7-5a96-42e2-9807-38a6742f5ac3" Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.470778 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" event={"ID":"ef9d3954-d526-4bd6-b806-e0ddb31994d3","Type":"ContainerStarted","Data":"20548726e5615cb960c068e90e77e160478b4494815310eba43783f36e8bdbfc"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.482577 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" event={"ID":"d36f8a3a-450b-4824-826c-32283863f270","Type":"ContainerStarted","Data":"569029e0fd4a2f6aa65ac451f1f3d4a5476b333e478282f06d30a1607fe66476"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.482633 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.505213 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" podUID="d36f8a3a-450b-4824-826c-32283863f270" Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.522071 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" event={"ID":"64226f84-d145-459b-b326-acf7e87add59","Type":"ContainerStarted","Data":"ad6fffa2d68a7ba07b76a344ea7948ee2fcc7ccc9eaacacf990f68ffd28615ff"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.552668 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" event={"ID":"568f4286-ea33-4e00-ab87-cdc930373223","Type":"ContainerStarted","Data":"0306dfbc1ba8a358a8da59c5254af8e5c5583bc3bf339a3e5147a3029ad58615"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.552733 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" Dec 05 12:23:54 crc kubenswrapper[4807]: E1205 12:23:54.558839 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" podUID="568f4286-ea33-4e00-ab87-cdc930373223" Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.571858 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" event={"ID":"c383d37a-acb0-463c-9980-1d2d46e06291","Type":"ContainerStarted","Data":"b8a780d5f1edcb18ac57f2922d72b25d7c55d182e7d6b9dca74cccf5b327899f"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.589923 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" event={"ID":"7cd89edb-31e7-4c50-ba4d-c86f85fce855","Type":"ContainerStarted","Data":"a2af8bf7382b7661133e63d9cc2da1d3a6da10c915bebff45ed375c169541ba1"} Dec 05 12:23:54 crc kubenswrapper[4807]: I1205 12:23:54.615364 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" event={"ID":"2f6ae961-a4b7-4d96-8669-2b0c1653ddc3","Type":"ContainerStarted","Data":"9f0212d30472b8ae5f7da2aba59adb61923b670cbc5383f47379df0e65a56eea"} Dec 05 12:23:55 crc kubenswrapper[4807]: E1205 12:23:55.633816 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" podUID="d36f8a3a-450b-4824-826c-32283863f270" Dec 05 12:23:55 crc kubenswrapper[4807]: E1205 12:23:55.634287 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" podUID="568f4286-ea33-4e00-ab87-cdc930373223" Dec 05 12:23:55 crc kubenswrapper[4807]: E1205 12:23:55.634433 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" podUID="d2b94a74-f984-4360-8062-e42b82a7c401" Dec 05 12:23:55 crc kubenswrapper[4807]: E1205 12:23:55.634506 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" podUID="8817b3f7-5a96-42e2-9807-38a6742f5ac3" Dec 05 12:24:07 crc kubenswrapper[4807]: I1205 12:24:07.348979 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" Dec 05 12:24:07 crc kubenswrapper[4807]: I1205 12:24:07.368255 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" Dec 05 12:24:07 crc kubenswrapper[4807]: I1205 12:24:07.848154 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" Dec 05 12:24:07 crc kubenswrapper[4807]: I1205 12:24:07.936635 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" Dec 05 12:24:09 crc kubenswrapper[4807]: I1205 12:24:09.264910 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:24:09 crc kubenswrapper[4807]: I1205 12:24:09.271575 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a8a470d-bdcf-4d64-b805-c996e88384ae-cert\") pod \"infra-operator-controller-manager-57548d458d-v6m8x\" (UID: \"4a8a470d-bdcf-4d64-b805-c996e88384ae\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:24:09 crc kubenswrapper[4807]: I1205 12:24:09.410867 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:24:09 crc kubenswrapper[4807]: I1205 12:24:09.771663 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:24:09 crc kubenswrapper[4807]: I1205 12:24:09.776375 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8\" (UID: \"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:24:09 crc kubenswrapper[4807]: I1205 12:24:09.897334 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:24:10 crc kubenswrapper[4807]: I1205 12:24:10.278487 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:24:10 crc kubenswrapper[4807]: I1205 12:24:10.279147 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:24:10 crc kubenswrapper[4807]: I1205 12:24:10.282096 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-metrics-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:24:10 crc kubenswrapper[4807]: I1205 12:24:10.283510 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bc48b693-8a75-4dea-a3cf-3dd053a8a500-webhook-certs\") pod \"openstack-operator-controller-manager-fcb8bd8db-sbftd\" (UID: \"bc48b693-8a75-4dea-a3cf-3dd053a8a500\") " pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:24:10 crc kubenswrapper[4807]: I1205 12:24:10.310849 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:24:15 crc kubenswrapper[4807]: E1205 12:24:15.976394 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 05 12:24:15 crc kubenswrapper[4807]: E1205 12:24:15.977080 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66lz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-t8bz9_openstack-operators(49f023c0-dfc7-4035-ae05-df3e6913e581): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 12:24:16 crc kubenswrapper[4807]: E1205 12:24:16.028664 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 12:24:16 crc kubenswrapper[4807]: E1205 12:24:16.029076 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5fdsc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-j26nl_openstack-operators(ef9d3954-d526-4bd6-b806-e0ddb31994d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 12:24:16 crc kubenswrapper[4807]: E1205 12:24:16.030615 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" podUID="ef9d3954-d526-4bd6-b806-e0ddb31994d3" Dec 05 12:24:16 crc kubenswrapper[4807]: I1205 12:24:16.790083 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" Dec 05 12:24:16 crc kubenswrapper[4807]: E1205 12:24:16.791462 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" podUID="ef9d3954-d526-4bd6-b806-e0ddb31994d3" Dec 05 12:24:16 crc kubenswrapper[4807]: I1205 12:24:16.794162 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" Dec 05 12:24:17 crc kubenswrapper[4807]: E1205 12:24:17.161657 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 12:24:17 crc kubenswrapper[4807]: E1205 12:24:17.161844 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jsqnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-vl87h_openstack-operators(64226f84-d145-459b-b326-acf7e87add59): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 12:24:17 crc kubenswrapper[4807]: E1205 12:24:17.163086 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" podUID="64226f84-d145-459b-b326-acf7e87add59" Dec 05 12:24:17 crc kubenswrapper[4807]: E1205 12:24:17.693213 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 05 12:24:17 crc kubenswrapper[4807]: E1205 12:24:17.693463 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6t7h8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-r2q7v_openstack-operators(87b435a4-0ca7-4a37-9736-40ff511bfb43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 12:24:17 crc kubenswrapper[4807]: E1205 12:24:17.695070 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" podUID="87b435a4-0ca7-4a37-9736-40ff511bfb43" Dec 05 12:24:17 crc kubenswrapper[4807]: I1205 12:24:17.816962 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" Dec 05 12:24:17 crc kubenswrapper[4807]: I1205 12:24:17.820953 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.265594 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x"] Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.277202 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd"] Dec 05 12:24:18 crc kubenswrapper[4807]: E1205 12:24:18.344474 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" podUID="49f023c0-dfc7-4035-ae05-df3e6913e581" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.378868 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8"] Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.862099 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" event={"ID":"49f023c0-dfc7-4035-ae05-df3e6913e581","Type":"ContainerStarted","Data":"0a5cb4ae1a0724e18b668d06ee1cc0392ab6131dd1010a9c0715ee94834fb271"} Dec 05 12:24:18 crc kubenswrapper[4807]: E1205 12:24:18.867087 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" podUID="49f023c0-dfc7-4035-ae05-df3e6913e581" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.882875 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" event={"ID":"7cd89edb-31e7-4c50-ba4d-c86f85fce855","Type":"ContainerStarted","Data":"117f1ae4c2f19a98782528338f4a08cf598f83269d793456fa01c2e9d0e73e47"} Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.883345 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.885717 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.897560 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" event={"ID":"022ffaeb-bad6-48ee-be09-78fc4c515a99","Type":"ContainerStarted","Data":"f26dcc83f5620e8a386b58ea797f8f5c9028d9e7205ae7c7d1433e3aec430015"} Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.897775 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.906807 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.924334 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" event={"ID":"568f4286-ea33-4e00-ab87-cdc930373223","Type":"ContainerStarted","Data":"9a6b981c1bde6a200f37f2e0294a2dbdf44d89a0b3a684c188acc1358a81e507"} Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.931150 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-blrcs" podStartSLOduration=3.226974014 podStartE2EDuration="41.931126413s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.176438886 +0000 UTC m=+1048.670302155" lastFinishedPulling="2025-12-05 12:24:17.880591285 +0000 UTC m=+1087.374454554" observedRunningTime="2025-12-05 12:24:18.921944761 +0000 UTC m=+1088.415808030" watchObservedRunningTime="2025-12-05 12:24:18.931126413 +0000 UTC m=+1088.424989682" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.943161 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ng4kv" podStartSLOduration=3.957878663 podStartE2EDuration="42.943141566s" podCreationTimestamp="2025-12-05 12:23:36 +0000 UTC" firstStartedPulling="2025-12-05 12:23:38.914703572 +0000 UTC m=+1048.408566841" lastFinishedPulling="2025-12-05 12:24:17.899966475 +0000 UTC m=+1087.393829744" observedRunningTime="2025-12-05 12:24:18.942158791 +0000 UTC m=+1088.436022070" watchObservedRunningTime="2025-12-05 12:24:18.943141566 +0000 UTC m=+1088.437004835" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.946441 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" event={"ID":"c383d37a-acb0-463c-9980-1d2d46e06291","Type":"ContainerStarted","Data":"d64906fe7c6a197d49be0633a28d9e25f33bd7e99a0d0ecee04029fb7e3a036b"} Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.947405 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.954173 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" event={"ID":"4a8a470d-bdcf-4d64-b805-c996e88384ae","Type":"ContainerStarted","Data":"36811479aabec49b699b839947de9324eb44326e9fc4ae75ae8e2e98e85db105"} Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.954272 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" Dec 05 12:24:18 crc kubenswrapper[4807]: I1205 12:24:18.983272 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" event={"ID":"63696f3a-63f2-475e-b5e1-9ed44bc6eb39","Type":"ContainerStarted","Data":"5aecb9716f82bf328147eebdc0a589da98880298fbe4848154c305c9f4795ffe"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.001911 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rzhvm" podStartSLOduration=3.950002155 podStartE2EDuration="42.001881431s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.895958929 +0000 UTC m=+1049.389822198" lastFinishedPulling="2025-12-05 12:24:17.947838205 +0000 UTC m=+1087.441701474" observedRunningTime="2025-12-05 12:24:18.992173145 +0000 UTC m=+1088.486036414" watchObservedRunningTime="2025-12-05 12:24:19.001881431 +0000 UTC m=+1088.495744720" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.027179 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-knvh4" podStartSLOduration=3.318319022 podStartE2EDuration="42.02716067s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.162385691 +0000 UTC m=+1048.656248960" lastFinishedPulling="2025-12-05 12:24:17.871227339 +0000 UTC m=+1087.365090608" observedRunningTime="2025-12-05 12:24:19.02636208 +0000 UTC m=+1088.520225349" watchObservedRunningTime="2025-12-05 12:24:19.02716067 +0000 UTC m=+1088.521023939" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.032591 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" event={"ID":"2f6ae961-a4b7-4d96-8669-2b0c1653ddc3","Type":"ContainerStarted","Data":"64407ba6ab134533e36fb19bc7c5d4624fb48f0911e3ed09dafdbd1e3b824c18"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.033887 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.057054 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" event={"ID":"bc48b693-8a75-4dea-a3cf-3dd053a8a500","Type":"ContainerStarted","Data":"52829e7749998c1131419089a89c1cab7dff4874b5ff31d68f392786516fb40c"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.057108 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" event={"ID":"bc48b693-8a75-4dea-a3cf-3dd053a8a500","Type":"ContainerStarted","Data":"5a3b96003dc4e9ed7740164300ecd841c215034975808e04d8ff2916b9d30a1d"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.057814 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.068104 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.077681 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nclwx" podStartSLOduration=4.085700825 podStartE2EDuration="42.077664496s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.813356532 +0000 UTC m=+1049.307219801" lastFinishedPulling="2025-12-05 12:24:17.805320203 +0000 UTC m=+1087.299183472" observedRunningTime="2025-12-05 12:24:19.075661706 +0000 UTC m=+1088.569524975" watchObservedRunningTime="2025-12-05 12:24:19.077664496 +0000 UTC m=+1088.571527765" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.092509 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" event={"ID":"65950bcd-6c98-49b5-b25f-f58cd0ee72bd","Type":"ContainerStarted","Data":"5124476d0effed715bd08b4947508a374f23e486e14c9609c332e3e9628146b6"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.093643 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.103909 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.115929 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" event={"ID":"85952cd3-56c3-49d2-9f15-cd1f872a6c7c","Type":"ContainerStarted","Data":"887c9e2c2deac6a2bc3166fc26efbad13b35a67605d8dc2e5d6319bd9ad967cd"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.166571 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" podStartSLOduration=42.166520892 podStartE2EDuration="42.166520892s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:24:19.157630206 +0000 UTC m=+1088.651493465" watchObservedRunningTime="2025-12-05 12:24:19.166520892 +0000 UTC m=+1088.660384161" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.201109 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" event={"ID":"5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86","Type":"ContainerStarted","Data":"53540d56676b53f577fb07d50366de3ad17645ff31b6fe6f4da121fed6125d39"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.202907 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.223245 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.295006 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-5qfr7" podStartSLOduration=4.405060955 podStartE2EDuration="42.294948926s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:40.014900505 +0000 UTC m=+1049.508763774" lastFinishedPulling="2025-12-05 12:24:17.904788486 +0000 UTC m=+1087.398651745" observedRunningTime="2025-12-05 12:24:19.200834628 +0000 UTC m=+1088.694697897" watchObservedRunningTime="2025-12-05 12:24:19.294948926 +0000 UTC m=+1088.788812195" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.366626 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.366693 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.366706 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" event={"ID":"64763a6c-60c1-428e-baf9-040111b0057e","Type":"ContainerStarted","Data":"3f9fa11eeb03f0aef00d25729261494a528a0dd4df6ed19e5ece2435cc1160cb"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.368188 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" event={"ID":"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a","Type":"ContainerStarted","Data":"4c39e465696635ec76066bc8c33266f934284cf65f6b0c9eab832d394eefcaef"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.372814 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" event={"ID":"5be4b980-a6ab-4499-87ec-0aff47a7a917","Type":"ContainerStarted","Data":"ed1ba85896ff830f3884ad742a83d716065436115e8a53eeb6cd1a1c61c75013"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.373785 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.375444 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7pszd" podStartSLOduration=4.518938133 podStartE2EDuration="42.375426061s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:40.01471179 +0000 UTC m=+1049.508575059" lastFinishedPulling="2025-12-05 12:24:17.871199718 +0000 UTC m=+1087.365062987" observedRunningTime="2025-12-05 12:24:19.241119016 +0000 UTC m=+1088.734982315" watchObservedRunningTime="2025-12-05 12:24:19.375426061 +0000 UTC m=+1088.869289330" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.387423 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.411166 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" event={"ID":"3e3e2978-5f36-49af-ac8f-3ca92ada318e","Type":"ContainerStarted","Data":"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.411807 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.421544 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jzdm6" podStartSLOduration=4.483941457 podStartE2EDuration="42.421505345s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.821816695 +0000 UTC m=+1049.315679964" lastFinishedPulling="2025-12-05 12:24:17.759380573 +0000 UTC m=+1087.253243852" observedRunningTime="2025-12-05 12:24:19.389393444 +0000 UTC m=+1088.883256743" watchObservedRunningTime="2025-12-05 12:24:19.421505345 +0000 UTC m=+1088.915368624" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.425638 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" event={"ID":"64226f84-d145-459b-b326-acf7e87add59","Type":"ContainerStarted","Data":"a0daf827da5d21d507995748a24c107cb104f33d59b679ec940c99b10606919d"} Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.450373 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-g2mx8" podStartSLOduration=4.745496957 podStartE2EDuration="42.450350943s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.999582408 +0000 UTC m=+1049.493445677" lastFinishedPulling="2025-12-05 12:24:17.704436394 +0000 UTC m=+1087.198299663" observedRunningTime="2025-12-05 12:24:19.418267753 +0000 UTC m=+1088.912131032" watchObservedRunningTime="2025-12-05 12:24:19.450350943 +0000 UTC m=+1088.944214213" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.506592 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" podStartSLOduration=25.258202321 podStartE2EDuration="42.506564704s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:40.063823481 +0000 UTC m=+1049.557686750" lastFinishedPulling="2025-12-05 12:23:57.312185864 +0000 UTC m=+1066.806049133" observedRunningTime="2025-12-05 12:24:19.503017245 +0000 UTC m=+1088.996880534" watchObservedRunningTime="2025-12-05 12:24:19.506564704 +0000 UTC m=+1089.000427973" Dec 05 12:24:19 crc kubenswrapper[4807]: I1205 12:24:19.542513 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-vl87h" podStartSLOduration=29.060363953 podStartE2EDuration="42.542494592s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.709420875 +0000 UTC m=+1049.203284134" lastFinishedPulling="2025-12-05 12:23:53.191551504 +0000 UTC m=+1062.685414773" observedRunningTime="2025-12-05 12:24:19.540350438 +0000 UTC m=+1089.034213707" watchObservedRunningTime="2025-12-05 12:24:19.542494592 +0000 UTC m=+1089.036357861" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.438423 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" event={"ID":"d2b94a74-f984-4360-8062-e42b82a7c401","Type":"ContainerStarted","Data":"3b686a4063d822ea8a05f312fade559dd976cf533b8a562bfe41ed04e412e2e2"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.444451 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" event={"ID":"ef9d3954-d526-4bd6-b806-e0ddb31994d3","Type":"ContainerStarted","Data":"465d3f20f738f92ad8496b60cfdd7e7ff57d70d78e0c2f665d98275341fd7c95"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.446933 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" event={"ID":"3e3e2978-5f36-49af-ac8f-3ca92ada318e","Type":"ContainerStarted","Data":"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.452971 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" event={"ID":"d36f8a3a-450b-4824-826c-32283863f270","Type":"ContainerStarted","Data":"0c3137043b76e5cdddd1fa2605bfc64a54f7b665ada02f2a0bd50c1dc5411898"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.457550 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" event={"ID":"85952cd3-56c3-49d2-9f15-cd1f872a6c7c","Type":"ContainerStarted","Data":"e24f7dd54ffa281a4d836f09f693923b0eb92c8c74f41482b86a6a8ecad535de"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.457631 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.468256 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" event={"ID":"336d4305-fc09-4622-b6fd-ebf4fe266ad2","Type":"ContainerStarted","Data":"aae340a25afa1517bacfa1c29e17d103eae3c4dbcc400c598a77ca2520d66f82"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.470674 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.472567 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-zjzsd" podStartSLOduration=5.462482705 podStartE2EDuration="44.472552635s" podCreationTimestamp="2025-12-05 12:23:36 +0000 UTC" firstStartedPulling="2025-12-05 12:23:38.889824663 +0000 UTC m=+1048.383687942" lastFinishedPulling="2025-12-05 12:24:17.899894603 +0000 UTC m=+1087.393757872" observedRunningTime="2025-12-05 12:24:20.470873723 +0000 UTC m=+1089.964736992" watchObservedRunningTime="2025-12-05 12:24:20.472552635 +0000 UTC m=+1089.966415904" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.480568 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.481242 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" event={"ID":"8817b3f7-5a96-42e2-9807-38a6742f5ac3","Type":"ContainerStarted","Data":"717276cd6e9ba85f5d68f100a36e2bceb041c63e01a8b5dbbef27cf705b93af8"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.485463 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" event={"ID":"63696f3a-63f2-475e-b5e1-9ed44bc6eb39","Type":"ContainerStarted","Data":"5cae0ee9c7dd2fe6ae2a47550487dffe2bf0aeb60f87c2f3bddb5afc947d7d62"} Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.509294 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-j26nl" podStartSLOduration=30.210306223 podStartE2EDuration="43.509273664s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.832061884 +0000 UTC m=+1049.325925163" lastFinishedPulling="2025-12-05 12:23:53.131029335 +0000 UTC m=+1062.624892604" observedRunningTime="2025-12-05 12:24:20.490642032 +0000 UTC m=+1089.984505311" watchObservedRunningTime="2025-12-05 12:24:20.509273664 +0000 UTC m=+1090.003136933" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.513952 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" podStartSLOduration=5.859604762 podStartE2EDuration="43.513932741s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:40.042765219 +0000 UTC m=+1049.536628478" lastFinishedPulling="2025-12-05 12:24:17.697093168 +0000 UTC m=+1087.190956457" observedRunningTime="2025-12-05 12:24:20.508290038 +0000 UTC m=+1090.002153307" watchObservedRunningTime="2025-12-05 12:24:20.513932741 +0000 UTC m=+1090.007796010" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.555084 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-t82v9" podStartSLOduration=4.667224728 podStartE2EDuration="43.5550651s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.145944135 +0000 UTC m=+1048.639807394" lastFinishedPulling="2025-12-05 12:24:18.033784497 +0000 UTC m=+1087.527647766" observedRunningTime="2025-12-05 12:24:20.529209467 +0000 UTC m=+1090.023072746" watchObservedRunningTime="2025-12-05 12:24:20.5550651 +0000 UTC m=+1090.048928369" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.561394 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-99xgh" podStartSLOduration=5.532407173 podStartE2EDuration="43.56137913s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.850709946 +0000 UTC m=+1049.344573215" lastFinishedPulling="2025-12-05 12:24:17.879681903 +0000 UTC m=+1087.373545172" observedRunningTime="2025-12-05 12:24:20.552973828 +0000 UTC m=+1090.046837107" watchObservedRunningTime="2025-12-05 12:24:20.56137913 +0000 UTC m=+1090.055242399" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.572716 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g5tvq" podStartSLOduration=5.418143656 podStartE2EDuration="43.572697506s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:39.909813899 +0000 UTC m=+1049.403677178" lastFinishedPulling="2025-12-05 12:24:18.064367759 +0000 UTC m=+1087.558231028" observedRunningTime="2025-12-05 12:24:20.571377283 +0000 UTC m=+1090.065240552" watchObservedRunningTime="2025-12-05 12:24:20.572697506 +0000 UTC m=+1090.066560775" Dec 05 12:24:20 crc kubenswrapper[4807]: I1205 12:24:20.625562 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" podStartSLOduration=6.473366672 podStartE2EDuration="43.625538552s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:40.042947784 +0000 UTC m=+1049.536811053" lastFinishedPulling="2025-12-05 12:24:17.195119664 +0000 UTC m=+1086.688982933" observedRunningTime="2025-12-05 12:24:20.621652823 +0000 UTC m=+1090.115516102" watchObservedRunningTime="2025-12-05 12:24:20.625538552 +0000 UTC m=+1090.119401831" Dec 05 12:24:21 crc kubenswrapper[4807]: I1205 12:24:21.494503 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.501778 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" event={"ID":"4a8a470d-bdcf-4d64-b805-c996e88384ae","Type":"ContainerStarted","Data":"1af18d2d384151faf44808708d2d23756d1d8260ae65279998dbaa1d208b8af6"} Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.501820 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" event={"ID":"4a8a470d-bdcf-4d64-b805-c996e88384ae","Type":"ContainerStarted","Data":"6bd60f9902b8b57b95f00ba5d6cbb34d03eb910ea99f14f04cbf3b6f65005df8"} Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.502694 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.503747 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" event={"ID":"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a","Type":"ContainerStarted","Data":"8588552e53c8ab1cc79fee13bf756a7927ad4d43204971c6405bbefd66dbfeea"} Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.503866 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" event={"ID":"f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a","Type":"ContainerStarted","Data":"d0c91c6e17df12f094da0b1a1884cb9bee3ade4ff6d862a4189600e6d729e583"} Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.506807 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-mfzmp" Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.524369 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" podStartSLOduration=41.91715191 podStartE2EDuration="45.524345675s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:24:18.290068453 +0000 UTC m=+1087.783931722" lastFinishedPulling="2025-12-05 12:24:21.897262218 +0000 UTC m=+1091.391125487" observedRunningTime="2025-12-05 12:24:22.518194769 +0000 UTC m=+1092.012058038" watchObservedRunningTime="2025-12-05 12:24:22.524345675 +0000 UTC m=+1092.018208954" Dec 05 12:24:22 crc kubenswrapper[4807]: I1205 12:24:22.555502 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" podStartSLOduration=42.100339559 podStartE2EDuration="45.555480442s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:24:18.458313825 +0000 UTC m=+1087.952177094" lastFinishedPulling="2025-12-05 12:24:21.913454708 +0000 UTC m=+1091.407317977" observedRunningTime="2025-12-05 12:24:22.548904195 +0000 UTC m=+1092.042767474" watchObservedRunningTime="2025-12-05 12:24:22.555480442 +0000 UTC m=+1092.049343721" Dec 05 12:24:23 crc kubenswrapper[4807]: I1205 12:24:23.510666 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:24:28 crc kubenswrapper[4807]: I1205 12:24:28.354312 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-zbdbh" Dec 05 12:24:28 crc kubenswrapper[4807]: I1205 12:24:28.483815 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:24:29 crc kubenswrapper[4807]: I1205 12:24:29.417481 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-v6m8x" Dec 05 12:24:29 crc kubenswrapper[4807]: I1205 12:24:29.903650 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8" Dec 05 12:24:30 crc kubenswrapper[4807]: E1205 12:24:30.237606 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" podUID="87b435a4-0ca7-4a37-9736-40ff511bfb43" Dec 05 12:24:30 crc kubenswrapper[4807]: I1205 12:24:30.316883 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-fcb8bd8db-sbftd" Dec 05 12:24:33 crc kubenswrapper[4807]: E1205 12:24:33.236913 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" podUID="49f023c0-dfc7-4035-ae05-df3e6913e581" Dec 05 12:24:43 crc kubenswrapper[4807]: I1205 12:24:43.677141 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" event={"ID":"87b435a4-0ca7-4a37-9736-40ff511bfb43","Type":"ContainerStarted","Data":"9dfb4833508edb14b66a97d9018df2751362ed34762810c024c083a4a51f9c39"} Dec 05 12:24:43 crc kubenswrapper[4807]: I1205 12:24:43.696212 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r2q7v" podStartSLOduration=3.7027379480000002 podStartE2EDuration="1m6.696189498s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:40.043470306 +0000 UTC m=+1049.537333575" lastFinishedPulling="2025-12-05 12:24:43.036921856 +0000 UTC m=+1112.530785125" observedRunningTime="2025-12-05 12:24:43.689788589 +0000 UTC m=+1113.183651858" watchObservedRunningTime="2025-12-05 12:24:43.696189498 +0000 UTC m=+1113.190052797" Dec 05 12:24:45 crc kubenswrapper[4807]: I1205 12:24:45.692409 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" event={"ID":"49f023c0-dfc7-4035-ae05-df3e6913e581","Type":"ContainerStarted","Data":"a664fa63c41e750579b932758ad1c7cfe9a3717b31baa0917d539a1bbde6d979"} Dec 05 12:24:45 crc kubenswrapper[4807]: I1205 12:24:45.692942 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" Dec 05 12:24:45 crc kubenswrapper[4807]: I1205 12:24:45.712966 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" podStartSLOduration=4.131531922 podStartE2EDuration="1m8.712943304s" podCreationTimestamp="2025-12-05 12:23:37 +0000 UTC" firstStartedPulling="2025-12-05 12:23:40.05622681 +0000 UTC m=+1049.550090079" lastFinishedPulling="2025-12-05 12:24:44.637638192 +0000 UTC m=+1114.131501461" observedRunningTime="2025-12-05 12:24:45.708866274 +0000 UTC m=+1115.202729543" watchObservedRunningTime="2025-12-05 12:24:45.712943304 +0000 UTC m=+1115.206806573" Dec 05 12:24:58 crc kubenswrapper[4807]: I1205 12:24:58.052451 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-t8bz9" Dec 05 12:25:03 crc kubenswrapper[4807]: I1205 12:25:03.587775 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7"] Dec 05 12:25:03 crc kubenswrapper[4807]: I1205 12:25:03.588640 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" podUID="4e11a885-2cc7-43f4-9626-4aa22223946b" containerName="operator" containerID="cri-o://211fe2d123dcf47d2417ac38c8d2eff409d1f3ab7680aa964ade415089c5f56d" gracePeriod=10 Dec 05 12:25:03 crc kubenswrapper[4807]: I1205 12:25:03.860501 4807 generic.go:334] "Generic (PLEG): container finished" podID="4e11a885-2cc7-43f4-9626-4aa22223946b" containerID="211fe2d123dcf47d2417ac38c8d2eff409d1f3ab7680aa964ade415089c5f56d" exitCode=0 Dec 05 12:25:03 crc kubenswrapper[4807]: I1205 12:25:03.860590 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" event={"ID":"4e11a885-2cc7-43f4-9626-4aa22223946b","Type":"ContainerDied","Data":"211fe2d123dcf47d2417ac38c8d2eff409d1f3ab7680aa964ade415089c5f56d"} Dec 05 12:25:03 crc kubenswrapper[4807]: I1205 12:25:03.992922 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.106415 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbwzf\" (UniqueName: \"kubernetes.io/projected/4e11a885-2cc7-43f4-9626-4aa22223946b-kube-api-access-bbwzf\") pod \"4e11a885-2cc7-43f4-9626-4aa22223946b\" (UID: \"4e11a885-2cc7-43f4-9626-4aa22223946b\") " Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.111979 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e11a885-2cc7-43f4-9626-4aa22223946b-kube-api-access-bbwzf" (OuterVolumeSpecName: "kube-api-access-bbwzf") pod "4e11a885-2cc7-43f4-9626-4aa22223946b" (UID: "4e11a885-2cc7-43f4-9626-4aa22223946b"). InnerVolumeSpecName "kube-api-access-bbwzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.208746 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbwzf\" (UniqueName: \"kubernetes.io/projected/4e11a885-2cc7-43f4-9626-4aa22223946b-kube-api-access-bbwzf\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.872837 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" event={"ID":"4e11a885-2cc7-43f4-9626-4aa22223946b","Type":"ContainerDied","Data":"4ac787321eb8607c5ae411643484f10dfe7cb72ca1ec8ac29a1e8f9f0e41d2df"} Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.872904 4807 scope.go:117] "RemoveContainer" containerID="211fe2d123dcf47d2417ac38c8d2eff409d1f3ab7680aa964ade415089c5f56d" Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.873990 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7" Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.921662 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7"] Dec 05 12:25:04 crc kubenswrapper[4807]: I1205 12:25:04.929659 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-574d9f8c97-6nmd7"] Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.251214 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e11a885-2cc7-43f4-9626-4aa22223946b" path="/var/lib/kubelet/pods/4e11a885-2cc7-43f4-9626-4aa22223946b/volumes" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.377751 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck"] Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.378012 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="manager" containerID="cri-o://fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307" gracePeriod=10 Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.378053 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="kube-rbac-proxy" containerID="cri-o://5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d" gracePeriod=10 Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.814791 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.884166 4807 generic.go:334] "Generic (PLEG): container finished" podID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerID="5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d" exitCode=0 Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.884203 4807 generic.go:334] "Generic (PLEG): container finished" podID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerID="fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307" exitCode=0 Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.884227 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" event={"ID":"3e3e2978-5f36-49af-ac8f-3ca92ada318e","Type":"ContainerDied","Data":"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d"} Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.884264 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" event={"ID":"3e3e2978-5f36-49af-ac8f-3ca92ada318e","Type":"ContainerDied","Data":"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307"} Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.884280 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" event={"ID":"3e3e2978-5f36-49af-ac8f-3ca92ada318e","Type":"ContainerDied","Data":"49e00b4a435eb31a6b292328884746d2ca775cab1b3862edd0c1f06476b281b8"} Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.884302 4807 scope.go:117] "RemoveContainer" containerID="5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.884305 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.899716 4807 scope.go:117] "RemoveContainer" containerID="fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.915367 4807 scope.go:117] "RemoveContainer" containerID="5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d" Dec 05 12:25:05 crc kubenswrapper[4807]: E1205 12:25:05.915809 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d\": container with ID starting with 5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d not found: ID does not exist" containerID="5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.915860 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d"} err="failed to get container status \"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d\": rpc error: code = NotFound desc = could not find container \"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d\": container with ID starting with 5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d not found: ID does not exist" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.915899 4807 scope.go:117] "RemoveContainer" containerID="fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307" Dec 05 12:25:05 crc kubenswrapper[4807]: E1205 12:25:05.916282 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307\": container with ID starting with fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307 not found: ID does not exist" containerID="fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.916338 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307"} err="failed to get container status \"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307\": rpc error: code = NotFound desc = could not find container \"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307\": container with ID starting with fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307 not found: ID does not exist" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.916374 4807 scope.go:117] "RemoveContainer" containerID="5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.916783 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d"} err="failed to get container status \"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d\": rpc error: code = NotFound desc = could not find container \"5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d\": container with ID starting with 5b7e13209e7aca90200a2a91e3ea4843a6425de0f42def0ad4ce34ceef7f438d not found: ID does not exist" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.916814 4807 scope.go:117] "RemoveContainer" containerID="fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.917171 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307"} err="failed to get container status \"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307\": rpc error: code = NotFound desc = could not find container \"fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307\": container with ID starting with fd44b55b00ed3435904cd10bb4aac710135e673595928925f2c6de80ebfbc307 not found: ID does not exist" Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.940972 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbgcp\" (UniqueName: \"kubernetes.io/projected/3e3e2978-5f36-49af-ac8f-3ca92ada318e-kube-api-access-dbgcp\") pod \"3e3e2978-5f36-49af-ac8f-3ca92ada318e\" (UID: \"3e3e2978-5f36-49af-ac8f-3ca92ada318e\") " Dec 05 12:25:05 crc kubenswrapper[4807]: I1205 12:25:05.946488 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e3e2978-5f36-49af-ac8f-3ca92ada318e-kube-api-access-dbgcp" (OuterVolumeSpecName: "kube-api-access-dbgcp") pod "3e3e2978-5f36-49af-ac8f-3ca92ada318e" (UID: "3e3e2978-5f36-49af-ac8f-3ca92ada318e"). InnerVolumeSpecName "kube-api-access-dbgcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:25:06 crc kubenswrapper[4807]: I1205 12:25:06.043023 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbgcp\" (UniqueName: \"kubernetes.io/projected/3e3e2978-5f36-49af-ac8f-3ca92ada318e-kube-api-access-dbgcp\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:06 crc kubenswrapper[4807]: I1205 12:25:06.215999 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck"] Dec 05 12:25:06 crc kubenswrapper[4807]: I1205 12:25:06.224434 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9545865c-bdpck"] Dec 05 12:25:07 crc kubenswrapper[4807]: I1205 12:25:07.245635 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" path="/var/lib/kubelet/pods/3e3e2978-5f36-49af-ac8f-3ca92ada318e/volumes" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.838465 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-index-vx7jt"] Dec 05 12:25:09 crc kubenswrapper[4807]: E1205 12:25:09.838871 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="manager" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.838888 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="manager" Dec 05 12:25:09 crc kubenswrapper[4807]: E1205 12:25:09.838940 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="kube-rbac-proxy" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.838948 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="kube-rbac-proxy" Dec 05 12:25:09 crc kubenswrapper[4807]: E1205 12:25:09.838970 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e11a885-2cc7-43f4-9626-4aa22223946b" containerName="operator" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.838978 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e11a885-2cc7-43f4-9626-4aa22223946b" containerName="operator" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.839131 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e11a885-2cc7-43f4-9626-4aa22223946b" containerName="operator" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.839148 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="manager" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.839168 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3e2978-5f36-49af-ac8f-3ca92ada318e" containerName="kube-rbac-proxy" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.839774 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.841626 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-index-dockercfg-c6qs9" Dec 05 12:25:09 crc kubenswrapper[4807]: I1205 12:25:09.847118 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-index-vx7jt"] Dec 05 12:25:10 crc kubenswrapper[4807]: I1205 12:25:10.011882 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt9mm\" (UniqueName: \"kubernetes.io/projected/49dfc8b3-c6d0-4c3f-a056-f6e4180068bd-kube-api-access-vt9mm\") pod \"watcher-operator-index-vx7jt\" (UID: \"49dfc8b3-c6d0-4c3f-a056-f6e4180068bd\") " pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:10 crc kubenswrapper[4807]: I1205 12:25:10.114120 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt9mm\" (UniqueName: \"kubernetes.io/projected/49dfc8b3-c6d0-4c3f-a056-f6e4180068bd-kube-api-access-vt9mm\") pod \"watcher-operator-index-vx7jt\" (UID: \"49dfc8b3-c6d0-4c3f-a056-f6e4180068bd\") " pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:10 crc kubenswrapper[4807]: I1205 12:25:10.148999 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt9mm\" (UniqueName: \"kubernetes.io/projected/49dfc8b3-c6d0-4c3f-a056-f6e4180068bd-kube-api-access-vt9mm\") pod \"watcher-operator-index-vx7jt\" (UID: \"49dfc8b3-c6d0-4c3f-a056-f6e4180068bd\") " pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:10 crc kubenswrapper[4807]: I1205 12:25:10.161147 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:10 crc kubenswrapper[4807]: I1205 12:25:10.635487 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-index-vx7jt"] Dec 05 12:25:10 crc kubenswrapper[4807]: I1205 12:25:10.919925 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-vx7jt" event={"ID":"49dfc8b3-c6d0-4c3f-a056-f6e4180068bd","Type":"ContainerStarted","Data":"e382408854e27b365a3cb383fa82d8d541f289510a767f60a4d30075f13201ab"} Dec 05 12:25:12 crc kubenswrapper[4807]: I1205 12:25:12.941655 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-vx7jt" event={"ID":"49dfc8b3-c6d0-4c3f-a056-f6e4180068bd","Type":"ContainerStarted","Data":"ca9b668165e606adba2023a6c0f0cbbc4a3381115181b00e7efacde1c1bdd152"} Dec 05 12:25:12 crc kubenswrapper[4807]: I1205 12:25:12.957774 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-index-vx7jt" podStartSLOduration=2.70265191 podStartE2EDuration="3.957757755s" podCreationTimestamp="2025-12-05 12:25:09 +0000 UTC" firstStartedPulling="2025-12-05 12:25:10.637725344 +0000 UTC m=+1140.131588613" lastFinishedPulling="2025-12-05 12:25:11.892831189 +0000 UTC m=+1141.386694458" observedRunningTime="2025-12-05 12:25:12.956423652 +0000 UTC m=+1142.450286921" watchObservedRunningTime="2025-12-05 12:25:12.957757755 +0000 UTC m=+1142.451621024" Dec 05 12:25:20 crc kubenswrapper[4807]: I1205 12:25:20.161921 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:20 crc kubenswrapper[4807]: I1205 12:25:20.162304 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:20 crc kubenswrapper[4807]: I1205 12:25:20.188687 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:21 crc kubenswrapper[4807]: I1205 12:25:21.030697 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-index-vx7jt" Dec 05 12:25:22 crc kubenswrapper[4807]: I1205 12:25:22.465742 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:25:22 crc kubenswrapper[4807]: I1205 12:25:22.466087 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:25:22 crc kubenswrapper[4807]: I1205 12:25:22.905351 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq"] Dec 05 12:25:22 crc kubenswrapper[4807]: I1205 12:25:22.906820 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:22 crc kubenswrapper[4807]: I1205 12:25:22.909839 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-rsvqr" Dec 05 12:25:22 crc kubenswrapper[4807]: I1205 12:25:22.923930 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq"] Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.002176 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-util\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.002592 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmlcd\" (UniqueName: \"kubernetes.io/projected/652de9e9-c06e-4661-81f6-8dec66fa510f-kube-api-access-cmlcd\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.002768 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-bundle\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.103225 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-bundle\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.103576 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-util\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.103793 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-bundle\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.103957 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmlcd\" (UniqueName: \"kubernetes.io/projected/652de9e9-c06e-4661-81f6-8dec66fa510f-kube-api-access-cmlcd\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.104078 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-util\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.125375 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmlcd\" (UniqueName: \"kubernetes.io/projected/652de9e9-c06e-4661-81f6-8dec66fa510f-kube-api-access-cmlcd\") pod \"6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.225829 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:23 crc kubenswrapper[4807]: I1205 12:25:23.631949 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq"] Dec 05 12:25:24 crc kubenswrapper[4807]: I1205 12:25:24.016483 4807 generic.go:334] "Generic (PLEG): container finished" podID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerID="bd44640b87757e901965f37c9fa3114debb54cb3f103131f3d84ca9ad3967f17" exitCode=0 Dec 05 12:25:24 crc kubenswrapper[4807]: I1205 12:25:24.016768 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" event={"ID":"652de9e9-c06e-4661-81f6-8dec66fa510f","Type":"ContainerDied","Data":"bd44640b87757e901965f37c9fa3114debb54cb3f103131f3d84ca9ad3967f17"} Dec 05 12:25:24 crc kubenswrapper[4807]: I1205 12:25:24.016800 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" event={"ID":"652de9e9-c06e-4661-81f6-8dec66fa510f","Type":"ContainerStarted","Data":"7251917f7250ebaf83c8db9140f9c185834e1d4af7c8523ebc74d6e166d17bb1"} Dec 05 12:25:25 crc kubenswrapper[4807]: I1205 12:25:25.050827 4807 generic.go:334] "Generic (PLEG): container finished" podID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerID="847b8dcfc28db9e2489b6a7ee0b4bc8c34c10c6cba03db9d4bcca0ce642a04f1" exitCode=0 Dec 05 12:25:25 crc kubenswrapper[4807]: I1205 12:25:25.051124 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" event={"ID":"652de9e9-c06e-4661-81f6-8dec66fa510f","Type":"ContainerDied","Data":"847b8dcfc28db9e2489b6a7ee0b4bc8c34c10c6cba03db9d4bcca0ce642a04f1"} Dec 05 12:25:26 crc kubenswrapper[4807]: I1205 12:25:26.060012 4807 generic.go:334] "Generic (PLEG): container finished" podID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerID="41d3f0d097d5b220217e5a56689980dc5f9d8071f80b850818f6bc2b2c057218" exitCode=0 Dec 05 12:25:26 crc kubenswrapper[4807]: I1205 12:25:26.060247 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" event={"ID":"652de9e9-c06e-4661-81f6-8dec66fa510f","Type":"ContainerDied","Data":"41d3f0d097d5b220217e5a56689980dc5f9d8071f80b850818f6bc2b2c057218"} Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.412674 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.474920 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmlcd\" (UniqueName: \"kubernetes.io/projected/652de9e9-c06e-4661-81f6-8dec66fa510f-kube-api-access-cmlcd\") pod \"652de9e9-c06e-4661-81f6-8dec66fa510f\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.474991 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-bundle\") pod \"652de9e9-c06e-4661-81f6-8dec66fa510f\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.475109 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-util\") pod \"652de9e9-c06e-4661-81f6-8dec66fa510f\" (UID: \"652de9e9-c06e-4661-81f6-8dec66fa510f\") " Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.476085 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-bundle" (OuterVolumeSpecName: "bundle") pod "652de9e9-c06e-4661-81f6-8dec66fa510f" (UID: "652de9e9-c06e-4661-81f6-8dec66fa510f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.500761 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/652de9e9-c06e-4661-81f6-8dec66fa510f-kube-api-access-cmlcd" (OuterVolumeSpecName: "kube-api-access-cmlcd") pod "652de9e9-c06e-4661-81f6-8dec66fa510f" (UID: "652de9e9-c06e-4661-81f6-8dec66fa510f"). InnerVolumeSpecName "kube-api-access-cmlcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.509692 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-util" (OuterVolumeSpecName: "util") pod "652de9e9-c06e-4661-81f6-8dec66fa510f" (UID: "652de9e9-c06e-4661-81f6-8dec66fa510f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.576432 4807 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-util\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.576488 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmlcd\" (UniqueName: \"kubernetes.io/projected/652de9e9-c06e-4661-81f6-8dec66fa510f-kube-api-access-cmlcd\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:27 crc kubenswrapper[4807]: I1205 12:25:27.576501 4807 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/652de9e9-c06e-4661-81f6-8dec66fa510f-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:28 crc kubenswrapper[4807]: I1205 12:25:28.080356 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" event={"ID":"652de9e9-c06e-4661-81f6-8dec66fa510f","Type":"ContainerDied","Data":"7251917f7250ebaf83c8db9140f9c185834e1d4af7c8523ebc74d6e166d17bb1"} Dec 05 12:25:28 crc kubenswrapper[4807]: I1205 12:25:28.080634 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7251917f7250ebaf83c8db9140f9c185834e1d4af7c8523ebc74d6e166d17bb1" Dec 05 12:25:28 crc kubenswrapper[4807]: I1205 12:25:28.080461 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.755713 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc"] Dec 05 12:25:32 crc kubenswrapper[4807]: E1205 12:25:32.756515 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerName="pull" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.756544 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerName="pull" Dec 05 12:25:32 crc kubenswrapper[4807]: E1205 12:25:32.756565 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerName="util" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.756571 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerName="util" Dec 05 12:25:32 crc kubenswrapper[4807]: E1205 12:25:32.756585 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerName="extract" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.756592 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerName="extract" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.756766 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="652de9e9-c06e-4661-81f6-8dec66fa510f" containerName="extract" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.757291 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.760849 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-service-cert" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.763618 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-x59cb" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.767134 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc"] Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.859767 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-apiservice-cert\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.860108 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95cbh\" (UniqueName: \"kubernetes.io/projected/e931fc69-05f2-4131-ba37-ee2c24920a1e-kube-api-access-95cbh\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.860195 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-webhook-cert\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.961101 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-apiservice-cert\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.961161 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95cbh\" (UniqueName: \"kubernetes.io/projected/e931fc69-05f2-4131-ba37-ee2c24920a1e-kube-api-access-95cbh\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.961238 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-webhook-cert\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.967828 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-webhook-cert\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.968574 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-apiservice-cert\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:32 crc kubenswrapper[4807]: I1205 12:25:32.984925 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95cbh\" (UniqueName: \"kubernetes.io/projected/e931fc69-05f2-4131-ba37-ee2c24920a1e-kube-api-access-95cbh\") pod \"watcher-operator-controller-manager-6864bb9567-pkbzc\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:33 crc kubenswrapper[4807]: I1205 12:25:33.079400 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:33 crc kubenswrapper[4807]: I1205 12:25:33.556259 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc"] Dec 05 12:25:33 crc kubenswrapper[4807]: W1205 12:25:33.558331 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode931fc69_05f2_4131_ba37_ee2c24920a1e.slice/crio-1361874947b88b2b91671a69f9bfd6ff6ccf9cb09b92b586a0a6c037193687d7 WatchSource:0}: Error finding container 1361874947b88b2b91671a69f9bfd6ff6ccf9cb09b92b586a0a6c037193687d7: Status 404 returned error can't find the container with id 1361874947b88b2b91671a69f9bfd6ff6ccf9cb09b92b586a0a6c037193687d7 Dec 05 12:25:34 crc kubenswrapper[4807]: I1205 12:25:34.122193 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" event={"ID":"e931fc69-05f2-4131-ba37-ee2c24920a1e","Type":"ContainerStarted","Data":"6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c"} Dec 05 12:25:34 crc kubenswrapper[4807]: I1205 12:25:34.122503 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" event={"ID":"e931fc69-05f2-4131-ba37-ee2c24920a1e","Type":"ContainerStarted","Data":"1361874947b88b2b91671a69f9bfd6ff6ccf9cb09b92b586a0a6c037193687d7"} Dec 05 12:25:34 crc kubenswrapper[4807]: I1205 12:25:34.122519 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:34 crc kubenswrapper[4807]: I1205 12:25:34.143025 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" podStartSLOduration=2.143005826 podStartE2EDuration="2.143005826s" podCreationTimestamp="2025-12-05 12:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:25:34.140114824 +0000 UTC m=+1163.633978093" watchObservedRunningTime="2025-12-05 12:25:34.143005826 +0000 UTC m=+1163.636869095" Dec 05 12:25:43 crc kubenswrapper[4807]: I1205 12:25:43.083995 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:44 crc kubenswrapper[4807]: I1205 12:25:44.748877 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt"] Dec 05 12:25:44 crc kubenswrapper[4807]: I1205 12:25:44.750212 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:44 crc kubenswrapper[4807]: I1205 12:25:44.769065 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt"] Dec 05 12:25:44 crc kubenswrapper[4807]: I1205 12:25:44.920030 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae6e3e1b-30de-427e-9616-8d023b6b97c5-webhook-cert\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:44 crc kubenswrapper[4807]: I1205 12:25:44.920083 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae6e3e1b-30de-427e-9616-8d023b6b97c5-apiservice-cert\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:44 crc kubenswrapper[4807]: I1205 12:25:44.920158 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvll8\" (UniqueName: \"kubernetes.io/projected/ae6e3e1b-30de-427e-9616-8d023b6b97c5-kube-api-access-pvll8\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.021785 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvll8\" (UniqueName: \"kubernetes.io/projected/ae6e3e1b-30de-427e-9616-8d023b6b97c5-kube-api-access-pvll8\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.021870 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae6e3e1b-30de-427e-9616-8d023b6b97c5-webhook-cert\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.021899 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae6e3e1b-30de-427e-9616-8d023b6b97c5-apiservice-cert\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.031353 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae6e3e1b-30de-427e-9616-8d023b6b97c5-apiservice-cert\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.031353 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae6e3e1b-30de-427e-9616-8d023b6b97c5-webhook-cert\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.047049 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvll8\" (UniqueName: \"kubernetes.io/projected/ae6e3e1b-30de-427e-9616-8d023b6b97c5-kube-api-access-pvll8\") pod \"watcher-operator-controller-manager-7b5d5fdf58-88nkt\" (UID: \"ae6e3e1b-30de-427e-9616-8d023b6b97c5\") " pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.089972 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:45 crc kubenswrapper[4807]: I1205 12:25:45.301606 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt"] Dec 05 12:25:46 crc kubenswrapper[4807]: I1205 12:25:46.210497 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" event={"ID":"ae6e3e1b-30de-427e-9616-8d023b6b97c5","Type":"ContainerStarted","Data":"fadf82924a4ec403af0d542d6522b44e3c40935b08bc6301c1511e007b665b44"} Dec 05 12:25:46 crc kubenswrapper[4807]: I1205 12:25:46.210566 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" event={"ID":"ae6e3e1b-30de-427e-9616-8d023b6b97c5","Type":"ContainerStarted","Data":"9c24c299324b65d95a426f83edb23786829e32d671ad35e99a10bbaa22742bf5"} Dec 05 12:25:46 crc kubenswrapper[4807]: I1205 12:25:46.210674 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:46 crc kubenswrapper[4807]: I1205 12:25:46.229034 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" podStartSLOduration=2.229017143 podStartE2EDuration="2.229017143s" podCreationTimestamp="2025-12-05 12:25:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:25:46.224361699 +0000 UTC m=+1175.718224978" watchObservedRunningTime="2025-12-05 12:25:46.229017143 +0000 UTC m=+1175.722880412" Dec 05 12:25:52 crc kubenswrapper[4807]: I1205 12:25:52.465867 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:25:52 crc kubenswrapper[4807]: I1205 12:25:52.466369 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:25:55 crc kubenswrapper[4807]: I1205 12:25:55.094822 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7b5d5fdf58-88nkt" Dec 05 12:25:55 crc kubenswrapper[4807]: I1205 12:25:55.149559 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc"] Dec 05 12:25:55 crc kubenswrapper[4807]: I1205 12:25:55.149811 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" podUID="e931fc69-05f2-4131-ba37-ee2c24920a1e" containerName="manager" containerID="cri-o://6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c" gracePeriod=10 Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.104629 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.283391 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95cbh\" (UniqueName: \"kubernetes.io/projected/e931fc69-05f2-4131-ba37-ee2c24920a1e-kube-api-access-95cbh\") pod \"e931fc69-05f2-4131-ba37-ee2c24920a1e\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.284388 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-apiservice-cert\") pod \"e931fc69-05f2-4131-ba37-ee2c24920a1e\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.284815 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-webhook-cert\") pod \"e931fc69-05f2-4131-ba37-ee2c24920a1e\" (UID: \"e931fc69-05f2-4131-ba37-ee2c24920a1e\") " Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.291569 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "e931fc69-05f2-4131-ba37-ee2c24920a1e" (UID: "e931fc69-05f2-4131-ba37-ee2c24920a1e"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.291588 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "e931fc69-05f2-4131-ba37-ee2c24920a1e" (UID: "e931fc69-05f2-4131-ba37-ee2c24920a1e"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.291828 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e931fc69-05f2-4131-ba37-ee2c24920a1e-kube-api-access-95cbh" (OuterVolumeSpecName: "kube-api-access-95cbh") pod "e931fc69-05f2-4131-ba37-ee2c24920a1e" (UID: "e931fc69-05f2-4131-ba37-ee2c24920a1e"). InnerVolumeSpecName "kube-api-access-95cbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.295904 4807 generic.go:334] "Generic (PLEG): container finished" podID="e931fc69-05f2-4131-ba37-ee2c24920a1e" containerID="6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c" exitCode=0 Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.295944 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" event={"ID":"e931fc69-05f2-4131-ba37-ee2c24920a1e","Type":"ContainerDied","Data":"6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c"} Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.295974 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" event={"ID":"e931fc69-05f2-4131-ba37-ee2c24920a1e","Type":"ContainerDied","Data":"1361874947b88b2b91671a69f9bfd6ff6ccf9cb09b92b586a0a6c037193687d7"} Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.295994 4807 scope.go:117] "RemoveContainer" containerID="6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.296168 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.353016 4807 scope.go:117] "RemoveContainer" containerID="6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c" Dec 05 12:25:56 crc kubenswrapper[4807]: E1205 12:25:56.353769 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c\": container with ID starting with 6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c not found: ID does not exist" containerID="6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.353807 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c"} err="failed to get container status \"6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c\": rpc error: code = NotFound desc = could not find container \"6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c\": container with ID starting with 6b4489251f1fc2b84182e5998686a4d82e0c0e6c24078428488fa340b3cd325c not found: ID does not exist" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.354234 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc"] Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.358898 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6864bb9567-pkbzc"] Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.386737 4807 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.386941 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95cbh\" (UniqueName: \"kubernetes.io/projected/e931fc69-05f2-4131-ba37-ee2c24920a1e-kube-api-access-95cbh\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:56 crc kubenswrapper[4807]: I1205 12:25:56.387001 4807 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e931fc69-05f2-4131-ba37-ee2c24920a1e-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:57 crc kubenswrapper[4807]: I1205 12:25:57.244079 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e931fc69-05f2-4131-ba37-ee2c24920a1e" path="/var/lib/kubelet/pods/e931fc69-05f2-4131-ba37-ee2c24920a1e/volumes" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.866231 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/rabbitmq-server-0"] Dec 05 12:26:07 crc kubenswrapper[4807]: E1205 12:26:07.866922 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e931fc69-05f2-4131-ba37-ee2c24920a1e" containerName="manager" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.866939 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="e931fc69-05f2-4131-ba37-ee2c24920a1e" containerName="manager" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.867151 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="e931fc69-05f2-4131-ba37-ee2c24920a1e" containerName="manager" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.868068 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.870042 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-rabbitmq-svc" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.870172 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-config-data" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.870423 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-erlang-cookie" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.870451 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-plugins-conf" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.870464 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-server-conf" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.870903 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openshift-service-ca.crt" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.870959 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-default-user" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.871149 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-server-dockercfg-b8vvk" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.872887 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"kube-root-ca.crt" Dec 05 12:26:07 crc kubenswrapper[4807]: I1205 12:26:07.884637 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-server-0"] Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.045744 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046188 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046232 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046256 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046295 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046321 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74aad9a5-73e5-45a4-9087-c245511adcf0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046350 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-config-data\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046372 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74aad9a5-73e5-45a4-9087-c245511adcf0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046393 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046436 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.046459 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz585\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-kube-api-access-rz585\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.114992 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/rabbitmq-notifications-server-0"] Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.116237 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.118439 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-notifications-erlang-cookie" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.118479 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-notifications-server-dockercfg-sq4fh" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.118509 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-notifications-plugins-conf" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.119303 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-notifications-config-data" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.122059 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-notifications-default-user" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.122285 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-notifications-server-conf" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.122612 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-rabbitmq-notifications-svc" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.147444 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-notifications-server-0"] Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.147758 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74aad9a5-73e5-45a4-9087-c245511adcf0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.147809 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-config-data\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.147838 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74aad9a5-73e5-45a4-9087-c245511adcf0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.147866 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.147914 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.147939 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz585\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-kube-api-access-rz585\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.148001 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.148038 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.148078 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.148103 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.148143 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.148786 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.149086 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.149202 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-config-data\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.149696 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.150117 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74aad9a5-73e5-45a4-9087-c245511adcf0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.155048 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.163644 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.163989 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74aad9a5-73e5-45a4-9087-c245511adcf0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.164558 4807 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.164624 4807 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/42710079a02fe90b6f0313838fb91019369870293e806f58e05fef740cab28bc/globalmount\"" pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.173799 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74aad9a5-73e5-45a4-9087-c245511adcf0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.183520 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz585\" (UniqueName: \"kubernetes.io/projected/74aad9a5-73e5-45a4-9087-c245511adcf0-kube-api-access-rz585\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.210545 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ae9547e-3650-4abb-b0c4-1b185be521e1\") pod \"rabbitmq-server-0\" (UID: \"74aad9a5-73e5-45a4-9087-c245511adcf0\") " pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249273 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249337 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxmdj\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-kube-api-access-qxmdj\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249363 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249415 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249437 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249459 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249512 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249561 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249588 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249620 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.249642 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.350878 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.350934 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351008 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351038 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxmdj\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-kube-api-access-qxmdj\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351063 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351118 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351144 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351166 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351232 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351275 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351301 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.351782 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.353831 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.354207 4807 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.354235 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.354250 4807 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3c7f130a407ec1c1f507c7dbd990c7d67a75b814ff0285dcca356d5ba40fb007/globalmount\"" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.354333 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.356224 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.358935 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.365855 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.366176 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.370001 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.374671 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxmdj\" (UniqueName: \"kubernetes.io/projected/3c5fe773-7fd0-4c59-b0c9-1391e8944c1e-kube-api-access-qxmdj\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.408140 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-71e14f29-72d8-48b3-9cff-c8b2f5f32799\") pod \"rabbitmq-notifications-server-0\" (UID: \"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.434055 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.488808 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.901742 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-notifications-server-0"] Dec 05 12:26:08 crc kubenswrapper[4807]: W1205 12:26:08.903966 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c5fe773_7fd0_4c59_b0c9_1391e8944c1e.slice/crio-bfef3a8c2d3ffeb6bcee1c0399185e92e09ccf9f778f7a0492ba1f606c6e0de7 WatchSource:0}: Error finding container bfef3a8c2d3ffeb6bcee1c0399185e92e09ccf9f778f7a0492ba1f606c6e0de7: Status 404 returned error can't find the container with id bfef3a8c2d3ffeb6bcee1c0399185e92e09ccf9f778f7a0492ba1f606c6e0de7 Dec 05 12:26:08 crc kubenswrapper[4807]: I1205 12:26:08.993905 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-server-0"] Dec 05 12:26:09 crc kubenswrapper[4807]: W1205 12:26:09.017469 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74aad9a5_73e5_45a4_9087_c245511adcf0.slice/crio-25572217d012ea402196f489c8ce5dca4af5773d365d711caf9a9a5d0364de4a WatchSource:0}: Error finding container 25572217d012ea402196f489c8ce5dca4af5773d365d711caf9a9a5d0364de4a: Status 404 returned error can't find the container with id 25572217d012ea402196f489c8ce5dca4af5773d365d711caf9a9a5d0364de4a Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.341820 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/openstack-galera-0"] Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.343461 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.350175 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-galera-openstack-svc" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.351161 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openstack-config-data" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.351380 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openstack-scripts" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.352423 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"galera-openstack-dockercfg-szz6q" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.355891 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstack-galera-0"] Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.371915 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"combined-ca-bundle" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.389742 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e","Type":"ContainerStarted","Data":"bfef3a8c2d3ffeb6bcee1c0399185e92e09ccf9f778f7a0492ba1f606c6e0de7"} Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.391703 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"74aad9a5-73e5-45a4-9087-c245511adcf0","Type":"ContainerStarted","Data":"25572217d012ea402196f489c8ce5dca4af5773d365d711caf9a9a5d0364de4a"} Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471046 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-kolla-config\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471113 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471145 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3244125-122c-4cb6-b673-acc9bdb7e96f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471219 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3244125-122c-4cb6-b673-acc9bdb7e96f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471272 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3244125-122c-4cb6-b673-acc9bdb7e96f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471295 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l76hr\" (UniqueName: \"kubernetes.io/projected/d3244125-122c-4cb6-b673-acc9bdb7e96f-kube-api-access-l76hr\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471345 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e43177e3-1eb1-402d-9833-0325c14ec965\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e43177e3-1eb1-402d-9833-0325c14ec965\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.471489 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-config-data-default\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572545 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-kolla-config\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572596 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572627 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3244125-122c-4cb6-b673-acc9bdb7e96f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572685 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3244125-122c-4cb6-b673-acc9bdb7e96f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572719 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3244125-122c-4cb6-b673-acc9bdb7e96f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572749 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l76hr\" (UniqueName: \"kubernetes.io/projected/d3244125-122c-4cb6-b673-acc9bdb7e96f-kube-api-access-l76hr\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572789 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e43177e3-1eb1-402d-9833-0325c14ec965\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e43177e3-1eb1-402d-9833-0325c14ec965\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.572836 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-config-data-default\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.573994 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-config-data-default\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.574499 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-kolla-config\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.575801 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3244125-122c-4cb6-b673-acc9bdb7e96f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.577090 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3244125-122c-4cb6-b673-acc9bdb7e96f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.582502 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3244125-122c-4cb6-b673-acc9bdb7e96f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.594361 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3244125-122c-4cb6-b673-acc9bdb7e96f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.603258 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l76hr\" (UniqueName: \"kubernetes.io/projected/d3244125-122c-4cb6-b673-acc9bdb7e96f-kube-api-access-l76hr\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.618005 4807 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.618045 4807 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e43177e3-1eb1-402d-9833-0325c14ec965\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e43177e3-1eb1-402d-9833-0325c14ec965\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/87e477f902be6a4d1c0eda19c212fc7dd6ca0036eeadf928071261dab850b218/globalmount\"" pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.756306 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.757282 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.759856 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"memcached-config-data" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.759873 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"memcached-memcached-dockercfg-5rcwl" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.764368 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.770510 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-memcached-svc" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.877647 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.877709 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.877801 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxh7v\" (UniqueName: \"kubernetes.io/projected/c9b15b84-7f89-472d-8560-86d63c73b30b-kube-api-access-cxh7v\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.877826 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-kolla-config\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.877881 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-config-data\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.905733 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e43177e3-1eb1-402d-9833-0325c14ec965\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e43177e3-1eb1-402d-9833-0325c14ec965\") pod \"openstack-galera-0\" (UID: \"d3244125-122c-4cb6-b673-acc9bdb7e96f\") " pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.962485 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.979464 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-config-data\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.979666 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.979697 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.979732 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxh7v\" (UniqueName: \"kubernetes.io/projected/c9b15b84-7f89-472d-8560-86d63c73b30b-kube-api-access-cxh7v\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.979753 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-kolla-config\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.980370 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-kolla-config\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.989067 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:09 crc kubenswrapper[4807]: I1205 12:26:09.996851 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-config-data\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.007719 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.012556 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxh7v\" (UniqueName: \"kubernetes.io/projected/c9b15b84-7f89-472d-8560-86d63c73b30b-kube-api-access-cxh7v\") pod \"memcached-0\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.057329 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.058487 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.061549 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"telemetry-ceilometer-dockercfg-km6lq" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.077256 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.097205 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.184089 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx76q\" (UniqueName: \"kubernetes.io/projected/db9e65be-19f5-4180-9b0c-96f323e547a3-kube-api-access-sx76q\") pod \"kube-state-metrics-0\" (UID: \"db9e65be-19f5-4180-9b0c-96f323e547a3\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.290363 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx76q\" (UniqueName: \"kubernetes.io/projected/db9e65be-19f5-4180-9b0c-96f323e547a3-kube-api-access-sx76q\") pod \"kube-state-metrics-0\" (UID: \"db9e65be-19f5-4180-9b0c-96f323e547a3\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.321290 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx76q\" (UniqueName: \"kubernetes.io/projected/db9e65be-19f5-4180-9b0c-96f323e547a3-kube-api-access-sx76q\") pod \"kube-state-metrics-0\" (UID: \"db9e65be-19f5-4180-9b0c-96f323e547a3\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.380332 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.566813 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstack-galera-0"] Dec 05 12:26:10 crc kubenswrapper[4807]: W1205 12:26:10.590693 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3244125_122c_4cb6_b673_acc9bdb7e96f.slice/crio-c0618f5e57e0b4907375ef3438f609eaa5d9caeec65a2c46668a306c590e9fc0 WatchSource:0}: Error finding container c0618f5e57e0b4907375ef3438f609eaa5d9caeec65a2c46668a306c590e9fc0: Status 404 returned error can't find the container with id c0618f5e57e0b4907375ef3438f609eaa5d9caeec65a2c46668a306c590e9fc0 Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.742577 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:26:10 crc kubenswrapper[4807]: W1205 12:26:10.748120 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9b15b84_7f89_472d_8560_86d63c73b30b.slice/crio-9da67cc6fdb097f46a6767ff5e8975a4c46a47aa22951090af55740a81226114 WatchSource:0}: Error finding container 9da67cc6fdb097f46a6767ff5e8975a4c46a47aa22951090af55740a81226114: Status 404 returned error can't find the container with id 9da67cc6fdb097f46a6767ff5e8975a4c46a47aa22951090af55740a81226114 Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.849835 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/alertmanager-metric-storage-0"] Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.851298 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.854289 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"metric-storage-alertmanager-dockercfg-cnqj7" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.854343 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-tls-assets-0" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.854288 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-generated" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.854485 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-cluster-tls-config" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.854512 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-web-config" Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.865793 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/alertmanager-metric-storage-0"] Dec 05 12:26:10 crc kubenswrapper[4807]: I1205 12:26:10.930790 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:26:10 crc kubenswrapper[4807]: W1205 12:26:10.938155 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb9e65be_19f5_4180_9b0c_96f323e547a3.slice/crio-a0be75746c5b0aae5234f80c2ed309682ee0d3f3cf3293ea9c4564b26eec81d2 WatchSource:0}: Error finding container a0be75746c5b0aae5234f80c2ed309682ee0d3f3cf3293ea9c4564b26eec81d2: Status 404 returned error can't find the container with id a0be75746c5b0aae5234f80c2ed309682ee0d3f3cf3293ea9c4564b26eec81d2 Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.010197 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5q4t\" (UniqueName: \"kubernetes.io/projected/4f4aba46-1f95-43ed-8346-57da458a508f-kube-api-access-r5q4t\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.010275 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.010314 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4f4aba46-1f95-43ed-8346-57da458a508f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.010331 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4f4aba46-1f95-43ed-8346-57da458a508f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.010424 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.010456 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.010476 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/4f4aba46-1f95-43ed-8346-57da458a508f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.075898 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg"] Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.077120 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.080064 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-wpbbt" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.081240 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.086853 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg"] Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.112633 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4f4aba46-1f95-43ed-8346-57da458a508f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.112670 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4f4aba46-1f95-43ed-8346-57da458a508f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.112714 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.112750 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.112768 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/4f4aba46-1f95-43ed-8346-57da458a508f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.112826 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5q4t\" (UniqueName: \"kubernetes.io/projected/4f4aba46-1f95-43ed-8346-57da458a508f-kube-api-access-r5q4t\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.112847 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.114254 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/4f4aba46-1f95-43ed-8346-57da458a508f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.119817 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4f4aba46-1f95-43ed-8346-57da458a508f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.121098 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4f4aba46-1f95-43ed-8346-57da458a508f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.122279 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.126499 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.129982 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4f4aba46-1f95-43ed-8346-57da458a508f-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.130191 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5q4t\" (UniqueName: \"kubernetes.io/projected/4f4aba46-1f95-43ed-8346-57da458a508f-kube-api-access-r5q4t\") pod \"alertmanager-metric-storage-0\" (UID: \"4f4aba46-1f95-43ed-8346-57da458a508f\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.173022 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.217511 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9471f89-372d-4267-aad8-a850cbee6d37-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-t82lg\" (UID: \"d9471f89-372d-4267-aad8-a850cbee6d37\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.217587 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4hfk\" (UniqueName: \"kubernetes.io/projected/d9471f89-372d-4267-aad8-a850cbee6d37-kube-api-access-c4hfk\") pod \"observability-ui-dashboards-7d5fb4cbfb-t82lg\" (UID: \"d9471f89-372d-4267-aad8-a850cbee6d37\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.320341 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9471f89-372d-4267-aad8-a850cbee6d37-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-t82lg\" (UID: \"d9471f89-372d-4267-aad8-a850cbee6d37\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.320383 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4hfk\" (UniqueName: \"kubernetes.io/projected/d9471f89-372d-4267-aad8-a850cbee6d37-kube-api-access-c4hfk\") pod \"observability-ui-dashboards-7d5fb4cbfb-t82lg\" (UID: \"d9471f89-372d-4267-aad8-a850cbee6d37\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.335217 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 05 12:26:11 crc kubenswrapper[4807]: E1205 12:26:11.337984 4807 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Dec 05 12:26:11 crc kubenswrapper[4807]: E1205 12:26:11.338072 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9471f89-372d-4267-aad8-a850cbee6d37-serving-cert podName:d9471f89-372d-4267-aad8-a850cbee6d37 nodeName:}" failed. No retries permitted until 2025-12-05 12:26:11.838046093 +0000 UTC m=+1201.331909362 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d9471f89-372d-4267-aad8-a850cbee6d37-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-t82lg" (UID: "d9471f89-372d-4267-aad8-a850cbee6d37") : secret "observability-ui-dashboards" not found Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.352516 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.358875 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.366760 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-tls-assets-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.367002 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.367171 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.367450 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-web-config" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.367662 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"metric-storage-prometheus-dockercfg-rwh5f" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.370663 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.386665 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.394400 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4hfk\" (UniqueName: \"kubernetes.io/projected/d9471f89-372d-4267-aad8-a850cbee6d37-kube-api-access-c4hfk\") pod \"observability-ui-dashboards-7d5fb4cbfb-t82lg\" (UID: \"d9471f89-372d-4267-aad8-a850cbee6d37\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.442984 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"db9e65be-19f5-4180-9b0c-96f323e547a3","Type":"ContainerStarted","Data":"a0be75746c5b0aae5234f80c2ed309682ee0d3f3cf3293ea9c4564b26eec81d2"} Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.444407 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"c9b15b84-7f89-472d-8560-86d63c73b30b","Type":"ContainerStarted","Data":"9da67cc6fdb097f46a6767ff5e8975a4c46a47aa22951090af55740a81226114"} Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.445344 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"d3244125-122c-4cb6-b673-acc9bdb7e96f","Type":"ContainerStarted","Data":"c0618f5e57e0b4907375ef3438f609eaa5d9caeec65a2c46668a306c590e9fc0"} Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.544900 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2774d099-8fed-443a-9e19-5200c7d41f37-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.544989 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.545018 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2774d099-8fed-443a-9e19-5200c7d41f37-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.545054 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.545181 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.545277 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dffg6\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-kube-api-access-dffg6\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.545311 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.545387 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-config\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.577066 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-74bf4df49-tr9s6"] Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.578931 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.589563 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74bf4df49-tr9s6"] Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.654030 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-oauth-serving-cert\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.654153 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-console-config\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.654231 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.654283 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-service-ca\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.654419 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dffg6\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-kube-api-access-dffg6\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.654823 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655141 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-config\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655190 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gwbw\" (UniqueName: \"kubernetes.io/projected/11518510-6028-4ad1-ae71-6fc74f433657-kube-api-access-6gwbw\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655220 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/11518510-6028-4ad1-ae71-6fc74f433657-console-serving-cert\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655280 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2774d099-8fed-443a-9e19-5200c7d41f37-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655296 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/11518510-6028-4ad1-ae71-6fc74f433657-console-oauth-config\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655322 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-trusted-ca-bundle\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655342 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655360 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2774d099-8fed-443a-9e19-5200c7d41f37-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.655386 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.656351 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2774d099-8fed-443a-9e19-5200c7d41f37-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.658363 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.659497 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-config\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.660166 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.668845 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2774d099-8fed-443a-9e19-5200c7d41f37-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.670089 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.675472 4807 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.675503 4807 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/76233c9f8195f01138e9549fd668a524184de5a9328019d10824c06f86e8a5fa/globalmount\"" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.676004 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dffg6\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-kube-api-access-dffg6\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.734088 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.742403 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.757182 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/11518510-6028-4ad1-ae71-6fc74f433657-console-oauth-config\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.757242 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-trusted-ca-bundle\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.757286 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-oauth-serving-cert\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.757309 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-console-config\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.757346 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-service-ca\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.757406 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gwbw\" (UniqueName: \"kubernetes.io/projected/11518510-6028-4ad1-ae71-6fc74f433657-kube-api-access-6gwbw\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.757435 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/11518510-6028-4ad1-ae71-6fc74f433657-console-serving-cert\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.758934 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-oauth-serving-cert\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.759338 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-console-config\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.762747 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/11518510-6028-4ad1-ae71-6fc74f433657-console-oauth-config\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.767965 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-service-ca\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.768177 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11518510-6028-4ad1-ae71-6fc74f433657-trusted-ca-bundle\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.770843 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/11518510-6028-4ad1-ae71-6fc74f433657-console-serving-cert\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.775690 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gwbw\" (UniqueName: \"kubernetes.io/projected/11518510-6028-4ad1-ae71-6fc74f433657-kube-api-access-6gwbw\") pod \"console-74bf4df49-tr9s6\" (UID: \"11518510-6028-4ad1-ae71-6fc74f433657\") " pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.848927 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/alertmanager-metric-storage-0"] Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.859196 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9471f89-372d-4267-aad8-a850cbee6d37-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-t82lg\" (UID: \"d9471f89-372d-4267-aad8-a850cbee6d37\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.862356 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9471f89-372d-4267-aad8-a850cbee6d37-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-t82lg\" (UID: \"d9471f89-372d-4267-aad8-a850cbee6d37\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.910030 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:11 crc kubenswrapper[4807]: I1205 12:26:11.993172 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-wpbbt" Dec 05 12:26:12 crc kubenswrapper[4807]: I1205 12:26:12.006627 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" Dec 05 12:26:12 crc kubenswrapper[4807]: I1205 12:26:12.281162 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:26:12 crc kubenswrapper[4807]: W1205 12:26:12.322685 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2774d099_8fed_443a_9e19_5200c7d41f37.slice/crio-fdf719874db513c7b9a5c36fe2bbb605aa7cc00a26373972d85a18e03095f3dc WatchSource:0}: Error finding container fdf719874db513c7b9a5c36fe2bbb605aa7cc00a26373972d85a18e03095f3dc: Status 404 returned error can't find the container with id fdf719874db513c7b9a5c36fe2bbb605aa7cc00a26373972d85a18e03095f3dc Dec 05 12:26:12 crc kubenswrapper[4807]: I1205 12:26:12.507811 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"4f4aba46-1f95-43ed-8346-57da458a508f","Type":"ContainerStarted","Data":"d06172da7299cb102259d81f0145c82cf9192d5a28611986196ee77367d02e88"} Dec 05 12:26:12 crc kubenswrapper[4807]: I1205 12:26:12.509412 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerStarted","Data":"fdf719874db513c7b9a5c36fe2bbb605aa7cc00a26373972d85a18e03095f3dc"} Dec 05 12:26:12 crc kubenswrapper[4807]: I1205 12:26:12.570110 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg"] Dec 05 12:26:12 crc kubenswrapper[4807]: I1205 12:26:12.595797 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74bf4df49-tr9s6"] Dec 05 12:26:13 crc kubenswrapper[4807]: I1205 12:26:13.519839 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" event={"ID":"d9471f89-372d-4267-aad8-a850cbee6d37","Type":"ContainerStarted","Data":"87664fa7e945043e039bc7403667528fc73a52e9655b91690ef190954f3837d6"} Dec 05 12:26:13 crc kubenswrapper[4807]: I1205 12:26:13.522504 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74bf4df49-tr9s6" event={"ID":"11518510-6028-4ad1-ae71-6fc74f433657","Type":"ContainerStarted","Data":"882f215d9cbb873d11015b9b8dd01d81edf6a974960cef164e53a8d9ce44b982"} Dec 05 12:26:13 crc kubenswrapper[4807]: I1205 12:26:13.522589 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74bf4df49-tr9s6" event={"ID":"11518510-6028-4ad1-ae71-6fc74f433657","Type":"ContainerStarted","Data":"20dce130383f52513404efa6b8bde543d61c00d8a161acd46f402b93cda8fdb8"} Dec 05 12:26:13 crc kubenswrapper[4807]: I1205 12:26:13.546101 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-74bf4df49-tr9s6" podStartSLOduration=2.546083966 podStartE2EDuration="2.546083966s" podCreationTimestamp="2025-12-05 12:26:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:26:13.541914043 +0000 UTC m=+1203.035777362" watchObservedRunningTime="2025-12-05 12:26:13.546083966 +0000 UTC m=+1203.039947235" Dec 05 12:26:21 crc kubenswrapper[4807]: I1205 12:26:21.911898 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:21 crc kubenswrapper[4807]: I1205 12:26:21.912306 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:21 crc kubenswrapper[4807]: I1205 12:26:21.916992 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:22 crc kubenswrapper[4807]: I1205 12:26:22.465974 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:26:22 crc kubenswrapper[4807]: I1205 12:26:22.466041 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:26:22 crc kubenswrapper[4807]: I1205 12:26:22.466089 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:26:22 crc kubenswrapper[4807]: I1205 12:26:22.466760 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b47bace47d16c688e6065026bd16a58454816df8e9f3811211e0c6bea14f1f68"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:26:22 crc kubenswrapper[4807]: I1205 12:26:22.466810 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://b47bace47d16c688e6065026bd16a58454816df8e9f3811211e0c6bea14f1f68" gracePeriod=600 Dec 05 12:26:22 crc kubenswrapper[4807]: I1205 12:26:22.606512 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-74bf4df49-tr9s6" Dec 05 12:26:22 crc kubenswrapper[4807]: I1205 12:26:22.656408 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5f5844bdbf-xdd6b"] Dec 05 12:26:24 crc kubenswrapper[4807]: I1205 12:26:24.622005 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="b47bace47d16c688e6065026bd16a58454816df8e9f3811211e0c6bea14f1f68" exitCode=0 Dec 05 12:26:24 crc kubenswrapper[4807]: I1205 12:26:24.622064 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"b47bace47d16c688e6065026bd16a58454816df8e9f3811211e0c6bea14f1f68"} Dec 05 12:26:24 crc kubenswrapper[4807]: I1205 12:26:24.622616 4807 scope.go:117] "RemoveContainer" containerID="9d6e7e624977c9e3e4da4ace9e23b2a19ff1d1ccbc78d45f4be87d54cc1e3e37" Dec 05 12:26:28 crc kubenswrapper[4807]: I1205 12:26:28.657177 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"ba4c03761578e8e0b9faf772d0c622df752a9d6198091d0d8315216f608b690f"} Dec 05 12:26:29 crc kubenswrapper[4807]: I1205 12:26:29.668614 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" event={"ID":"d9471f89-372d-4267-aad8-a850cbee6d37","Type":"ContainerStarted","Data":"7762b915ca12bfc284437ce9e29f5bd863bc1c66b26b7aa82906618e0731fff0"} Dec 05 12:26:29 crc kubenswrapper[4807]: I1205 12:26:29.674237 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"d3244125-122c-4cb6-b673-acc9bdb7e96f","Type":"ContainerStarted","Data":"5bfc2811b189e2e0f0110812da43e95d4c0f8182c5fb2f81b9fdf8dba1284b0b"} Dec 05 12:26:29 crc kubenswrapper[4807]: I1205 12:26:29.696592 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-t82lg" podStartSLOduration=3.3525572439999998 podStartE2EDuration="18.696571061s" podCreationTimestamp="2025-12-05 12:26:11 +0000 UTC" firstStartedPulling="2025-12-05 12:26:12.660273216 +0000 UTC m=+1202.154136485" lastFinishedPulling="2025-12-05 12:26:28.004287043 +0000 UTC m=+1217.498150302" observedRunningTime="2025-12-05 12:26:29.689365933 +0000 UTC m=+1219.183229192" watchObservedRunningTime="2025-12-05 12:26:29.696571061 +0000 UTC m=+1219.190434330" Dec 05 12:26:30 crc kubenswrapper[4807]: I1205 12:26:30.680938 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"c9b15b84-7f89-472d-8560-86d63c73b30b","Type":"ContainerStarted","Data":"a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184"} Dec 05 12:26:30 crc kubenswrapper[4807]: I1205 12:26:30.681260 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:30 crc kubenswrapper[4807]: I1205 12:26:30.682220 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e","Type":"ContainerStarted","Data":"4455aeb6c7803f06a7d02efc3b40e0744ac038528aab217bcbcaa6d51ade00d4"} Dec 05 12:26:30 crc kubenswrapper[4807]: I1205 12:26:30.699294 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/memcached-0" podStartSLOduration=4.76353551 podStartE2EDuration="21.699280259s" podCreationTimestamp="2025-12-05 12:26:09 +0000 UTC" firstStartedPulling="2025-12-05 12:26:10.752302099 +0000 UTC m=+1200.246165368" lastFinishedPulling="2025-12-05 12:26:27.688046848 +0000 UTC m=+1217.181910117" observedRunningTime="2025-12-05 12:26:30.698826738 +0000 UTC m=+1220.192690007" watchObservedRunningTime="2025-12-05 12:26:30.699280259 +0000 UTC m=+1220.193143528" Dec 05 12:26:31 crc kubenswrapper[4807]: I1205 12:26:31.691254 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerStarted","Data":"718faddb357f7c505739b3d88f49e913cd19aac5c9d06ca8e971fcf68125f5c8"} Dec 05 12:26:31 crc kubenswrapper[4807]: I1205 12:26:31.693102 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"74aad9a5-73e5-45a4-9087-c245511adcf0","Type":"ContainerStarted","Data":"a644cbfb8bfd2f1cae8c0c01e4ca0d7a57038f0dc0ae6df6c4571def9d71add8"} Dec 05 12:26:31 crc kubenswrapper[4807]: I1205 12:26:31.694721 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"db9e65be-19f5-4180-9b0c-96f323e547a3","Type":"ContainerStarted","Data":"83172aa6171e6fc96bd9242f36ccaba8d837e325626ba800ca11880c4f62ca35"} Dec 05 12:26:31 crc kubenswrapper[4807]: I1205 12:26:31.695010 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:26:31 crc kubenswrapper[4807]: I1205 12:26:31.756623 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/kube-state-metrics-0" podStartSLOduration=3.391991358 podStartE2EDuration="21.756599656s" podCreationTimestamp="2025-12-05 12:26:10 +0000 UTC" firstStartedPulling="2025-12-05 12:26:10.941158566 +0000 UTC m=+1200.435021835" lastFinishedPulling="2025-12-05 12:26:29.305766864 +0000 UTC m=+1218.799630133" observedRunningTime="2025-12-05 12:26:31.752975807 +0000 UTC m=+1221.246839076" watchObservedRunningTime="2025-12-05 12:26:31.756599656 +0000 UTC m=+1221.250462945" Dec 05 12:26:32 crc kubenswrapper[4807]: I1205 12:26:32.703897 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"4f4aba46-1f95-43ed-8346-57da458a508f","Type":"ContainerStarted","Data":"2b8a08aafa77cb70d4e37ecfa8713459d7c34fe3c439a312448c5f45aa46c67e"} Dec 05 12:26:35 crc kubenswrapper[4807]: I1205 12:26:35.099906 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/memcached-0" Dec 05 12:26:35 crc kubenswrapper[4807]: I1205 12:26:35.731541 4807 generic.go:334] "Generic (PLEG): container finished" podID="d3244125-122c-4cb6-b673-acc9bdb7e96f" containerID="5bfc2811b189e2e0f0110812da43e95d4c0f8182c5fb2f81b9fdf8dba1284b0b" exitCode=0 Dec 05 12:26:35 crc kubenswrapper[4807]: I1205 12:26:35.731581 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"d3244125-122c-4cb6-b673-acc9bdb7e96f","Type":"ContainerDied","Data":"5bfc2811b189e2e0f0110812da43e95d4c0f8182c5fb2f81b9fdf8dba1284b0b"} Dec 05 12:26:36 crc kubenswrapper[4807]: I1205 12:26:36.741654 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"d3244125-122c-4cb6-b673-acc9bdb7e96f","Type":"ContainerStarted","Data":"ce4f68f4132d69c8d2c811754940dca09f8e52c1e53d1c5538313a7c40db1f02"} Dec 05 12:26:36 crc kubenswrapper[4807]: I1205 12:26:36.764040 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/openstack-galera-0" podStartSLOduration=11.693219989 podStartE2EDuration="28.764021275s" podCreationTimestamp="2025-12-05 12:26:08 +0000 UTC" firstStartedPulling="2025-12-05 12:26:10.600128439 +0000 UTC m=+1200.093991708" lastFinishedPulling="2025-12-05 12:26:27.670929735 +0000 UTC m=+1217.164792994" observedRunningTime="2025-12-05 12:26:36.760025946 +0000 UTC m=+1226.253889215" watchObservedRunningTime="2025-12-05 12:26:36.764021275 +0000 UTC m=+1226.257884544" Dec 05 12:26:37 crc kubenswrapper[4807]: I1205 12:26:37.751476 4807 generic.go:334] "Generic (PLEG): container finished" podID="2774d099-8fed-443a-9e19-5200c7d41f37" containerID="718faddb357f7c505739b3d88f49e913cd19aac5c9d06ca8e971fcf68125f5c8" exitCode=0 Dec 05 12:26:37 crc kubenswrapper[4807]: I1205 12:26:37.751558 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerDied","Data":"718faddb357f7c505739b3d88f49e913cd19aac5c9d06ca8e971fcf68125f5c8"} Dec 05 12:26:38 crc kubenswrapper[4807]: I1205 12:26:38.761312 4807 generic.go:334] "Generic (PLEG): container finished" podID="4f4aba46-1f95-43ed-8346-57da458a508f" containerID="2b8a08aafa77cb70d4e37ecfa8713459d7c34fe3c439a312448c5f45aa46c67e" exitCode=0 Dec 05 12:26:38 crc kubenswrapper[4807]: I1205 12:26:38.761378 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"4f4aba46-1f95-43ed-8346-57da458a508f","Type":"ContainerDied","Data":"2b8a08aafa77cb70d4e37ecfa8713459d7c34fe3c439a312448c5f45aa46c67e"} Dec 05 12:26:39 crc kubenswrapper[4807]: I1205 12:26:39.963422 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:39 crc kubenswrapper[4807]: I1205 12:26:39.963840 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:40 crc kubenswrapper[4807]: I1205 12:26:40.385215 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:26:41 crc kubenswrapper[4807]: I1205 12:26:41.786078 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"4f4aba46-1f95-43ed-8346-57da458a508f","Type":"ContainerStarted","Data":"7c18f42dca9029054fa4928099df4f501e37c0c47053b7fa57ecb50902946634"} Dec 05 12:26:47 crc kubenswrapper[4807]: I1205 12:26:47.721818 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5f5844bdbf-xdd6b" podUID="ccee4f97-4c08-4396-af2e-c30ccae24fc7" containerName="console" containerID="cri-o://09483f27b026237683a2014ed52c9ff20960fa7cdf81da543af2169f92e22dc0" gracePeriod=15 Dec 05 12:26:48 crc kubenswrapper[4807]: I1205 12:26:48.844684 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5f5844bdbf-xdd6b_ccee4f97-4c08-4396-af2e-c30ccae24fc7/console/0.log" Dec 05 12:26:48 crc kubenswrapper[4807]: I1205 12:26:48.845217 4807 generic.go:334] "Generic (PLEG): container finished" podID="ccee4f97-4c08-4396-af2e-c30ccae24fc7" containerID="09483f27b026237683a2014ed52c9ff20960fa7cdf81da543af2169f92e22dc0" exitCode=2 Dec 05 12:26:48 crc kubenswrapper[4807]: I1205 12:26:48.845255 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f5844bdbf-xdd6b" event={"ID":"ccee4f97-4c08-4396-af2e-c30ccae24fc7","Type":"ContainerDied","Data":"09483f27b026237683a2014ed52c9ff20960fa7cdf81da543af2169f92e22dc0"} Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.251349 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.318719 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5f5844bdbf-xdd6b_ccee4f97-4c08-4396-af2e-c30ccae24fc7/console/0.log" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.318776 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.363696 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/openstack-galera-0" podUID="d3244125-122c-4cb6-b673-acc9bdb7e96f" containerName="galera" probeResult="failure" output=< Dec 05 12:26:49 crc kubenswrapper[4807]: wsrep_local_state_comment (Joined) differs from Synced Dec 05 12:26:49 crc kubenswrapper[4807]: > Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.401588 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-oauth-serving-cert\") pod \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.401661 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-service-ca\") pod \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.401727 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-config\") pod \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.401801 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-oauth-config\") pod \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.401851 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhktl\" (UniqueName: \"kubernetes.io/projected/ccee4f97-4c08-4396-af2e-c30ccae24fc7-kube-api-access-nhktl\") pod \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.401872 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-serving-cert\") pod \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.401902 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-trusted-ca-bundle\") pod \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\" (UID: \"ccee4f97-4c08-4396-af2e-c30ccae24fc7\") " Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.402337 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ccee4f97-4c08-4396-af2e-c30ccae24fc7" (UID: "ccee4f97-4c08-4396-af2e-c30ccae24fc7"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.402363 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-service-ca" (OuterVolumeSpecName: "service-ca") pod "ccee4f97-4c08-4396-af2e-c30ccae24fc7" (UID: "ccee4f97-4c08-4396-af2e-c30ccae24fc7"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.402560 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-config" (OuterVolumeSpecName: "console-config") pod "ccee4f97-4c08-4396-af2e-c30ccae24fc7" (UID: "ccee4f97-4c08-4396-af2e-c30ccae24fc7"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.402724 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ccee4f97-4c08-4396-af2e-c30ccae24fc7" (UID: "ccee4f97-4c08-4396-af2e-c30ccae24fc7"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.405229 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ccee4f97-4c08-4396-af2e-c30ccae24fc7" (UID: "ccee4f97-4c08-4396-af2e-c30ccae24fc7"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.405252 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ccee4f97-4c08-4396-af2e-c30ccae24fc7" (UID: "ccee4f97-4c08-4396-af2e-c30ccae24fc7"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.406544 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccee4f97-4c08-4396-af2e-c30ccae24fc7-kube-api-access-nhktl" (OuterVolumeSpecName: "kube-api-access-nhktl") pod "ccee4f97-4c08-4396-af2e-c30ccae24fc7" (UID: "ccee4f97-4c08-4396-af2e-c30ccae24fc7"). InnerVolumeSpecName "kube-api-access-nhktl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.503689 4807 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.503723 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhktl\" (UniqueName: \"kubernetes.io/projected/ccee4f97-4c08-4396-af2e-c30ccae24fc7-kube-api-access-nhktl\") on node \"crc\" DevicePath \"\"" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.503733 4807 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.503743 4807 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.503751 4807 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.503759 4807 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.503767 4807 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ccee4f97-4c08-4396-af2e-c30ccae24fc7-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.854467 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5f5844bdbf-xdd6b_ccee4f97-4c08-4396-af2e-c30ccae24fc7/console/0.log" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.854600 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f5844bdbf-xdd6b" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.854602 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f5844bdbf-xdd6b" event={"ID":"ccee4f97-4c08-4396-af2e-c30ccae24fc7","Type":"ContainerDied","Data":"2cc755b8d8ef52537ab7122546a5f607f146c0628e901841418f9ce986d6435e"} Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.854680 4807 scope.go:117] "RemoveContainer" containerID="09483f27b026237683a2014ed52c9ff20960fa7cdf81da543af2169f92e22dc0" Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.858729 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerStarted","Data":"a49e8a47a02b652f5cfcb766babe1fb648e69e2cf743cc589afa0d61bbcb0c4c"} Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.883650 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5f5844bdbf-xdd6b"] Dec 05 12:26:49 crc kubenswrapper[4807]: I1205 12:26:49.890991 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5f5844bdbf-xdd6b"] Dec 05 12:26:50 crc kubenswrapper[4807]: I1205 12:26:50.141646 4807 patch_prober.go:28] interesting pod/console-5f5844bdbf-xdd6b container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 12:26:50 crc kubenswrapper[4807]: I1205 12:26:50.141713 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5f5844bdbf-xdd6b" podUID="ccee4f97-4c08-4396-af2e-c30ccae24fc7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.49:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 12:26:50 crc kubenswrapper[4807]: I1205 12:26:50.352811 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/openstack-galera-0" Dec 05 12:26:51 crc kubenswrapper[4807]: I1205 12:26:51.245134 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccee4f97-4c08-4396-af2e-c30ccae24fc7" path="/var/lib/kubelet/pods/ccee4f97-4c08-4396-af2e-c30ccae24fc7/volumes" Dec 05 12:26:51 crc kubenswrapper[4807]: I1205 12:26:51.881658 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"4f4aba46-1f95-43ed-8346-57da458a508f","Type":"ContainerStarted","Data":"f81baa4d196edbc0ebfccb1f265916de47093dc69747b9d05709190c225b1fcf"} Dec 05 12:26:51 crc kubenswrapper[4807]: I1205 12:26:51.882722 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:51 crc kubenswrapper[4807]: I1205 12:26:51.885318 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Dec 05 12:26:51 crc kubenswrapper[4807]: I1205 12:26:51.906918 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/alertmanager-metric-storage-0" podStartSLOduration=12.701499586 podStartE2EDuration="41.906901631s" podCreationTimestamp="2025-12-05 12:26:10 +0000 UTC" firstStartedPulling="2025-12-05 12:26:11.882735493 +0000 UTC m=+1201.376598762" lastFinishedPulling="2025-12-05 12:26:41.088137538 +0000 UTC m=+1230.582000807" observedRunningTime="2025-12-05 12:26:51.904286267 +0000 UTC m=+1241.398149556" watchObservedRunningTime="2025-12-05 12:26:51.906901631 +0000 UTC m=+1241.400764900" Dec 05 12:26:52 crc kubenswrapper[4807]: I1205 12:26:52.893164 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerStarted","Data":"d621c6c67e81d7292a9c860401f5dfb4d82982e4ef73bfb1a374f496dc9a29e3"} Dec 05 12:26:57 crc kubenswrapper[4807]: I1205 12:26:57.938513 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerStarted","Data":"37d2a41cbe3f3de81af86995579cc202e4a63bdf6c2a82d79fca6fbea644f650"} Dec 05 12:26:57 crc kubenswrapper[4807]: I1205 12:26:57.963443 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/prometheus-metric-storage-0" podStartSLOduration=3.417845873 podStartE2EDuration="47.963424074s" podCreationTimestamp="2025-12-05 12:26:10 +0000 UTC" firstStartedPulling="2025-12-05 12:26:12.325076144 +0000 UTC m=+1201.818939413" lastFinishedPulling="2025-12-05 12:26:56.870654345 +0000 UTC m=+1246.364517614" observedRunningTime="2025-12-05 12:26:57.962154582 +0000 UTC m=+1247.456017851" watchObservedRunningTime="2025-12-05 12:26:57.963424074 +0000 UTC m=+1247.457287343" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.070770 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc"] Dec 05 12:27:00 crc kubenswrapper[4807]: E1205 12:27:00.071783 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccee4f97-4c08-4396-af2e-c30ccae24fc7" containerName="console" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.071802 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccee4f97-4c08-4396-af2e-c30ccae24fc7" containerName="console" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.072023 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccee4f97-4c08-4396-af2e-c30ccae24fc7" containerName="console" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.072793 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.076845 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-db-secret" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.081876 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-db-create-22lp6"] Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.083356 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.094769 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc"] Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.117416 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-create-22lp6"] Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.182292 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1bef399-bb15-4310-a951-08b0ca8ef58a-operator-scripts\") pod \"keystone-db-create-22lp6\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.182335 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87l5p\" (UniqueName: \"kubernetes.io/projected/d1bef399-bb15-4310-a951-08b0ca8ef58a-kube-api-access-87l5p\") pod \"keystone-db-create-22lp6\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.182381 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d95952-eefb-4b02-8480-aa8624398190-operator-scripts\") pod \"keystone-fd9f-account-create-update-5cpvc\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.182401 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvz5v\" (UniqueName: \"kubernetes.io/projected/a3d95952-eefb-4b02-8480-aa8624398190-kube-api-access-tvz5v\") pod \"keystone-fd9f-account-create-update-5cpvc\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.284091 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1bef399-bb15-4310-a951-08b0ca8ef58a-operator-scripts\") pod \"keystone-db-create-22lp6\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.284142 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87l5p\" (UniqueName: \"kubernetes.io/projected/d1bef399-bb15-4310-a951-08b0ca8ef58a-kube-api-access-87l5p\") pod \"keystone-db-create-22lp6\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.284186 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d95952-eefb-4b02-8480-aa8624398190-operator-scripts\") pod \"keystone-fd9f-account-create-update-5cpvc\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.284214 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvz5v\" (UniqueName: \"kubernetes.io/projected/a3d95952-eefb-4b02-8480-aa8624398190-kube-api-access-tvz5v\") pod \"keystone-fd9f-account-create-update-5cpvc\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.285307 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1bef399-bb15-4310-a951-08b0ca8ef58a-operator-scripts\") pod \"keystone-db-create-22lp6\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.285938 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d95952-eefb-4b02-8480-aa8624398190-operator-scripts\") pod \"keystone-fd9f-account-create-update-5cpvc\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.303678 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvz5v\" (UniqueName: \"kubernetes.io/projected/a3d95952-eefb-4b02-8480-aa8624398190-kube-api-access-tvz5v\") pod \"keystone-fd9f-account-create-update-5cpvc\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.304185 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87l5p\" (UniqueName: \"kubernetes.io/projected/d1bef399-bb15-4310-a951-08b0ca8ef58a-kube-api-access-87l5p\") pod \"keystone-db-create-22lp6\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.414501 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.422985 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.887077 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc"] Dec 05 12:27:00 crc kubenswrapper[4807]: W1205 12:27:00.889101 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3d95952_eefb_4b02_8480_aa8624398190.slice/crio-d7268cb060633909a08534442da1e772e28fe19ea6f03724716375bac2a58efc WatchSource:0}: Error finding container d7268cb060633909a08534442da1e772e28fe19ea6f03724716375bac2a58efc: Status 404 returned error can't find the container with id d7268cb060633909a08534442da1e772e28fe19ea6f03724716375bac2a58efc Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.961022 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-create-22lp6"] Dec 05 12:27:00 crc kubenswrapper[4807]: W1205 12:27:00.961912 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1bef399_bb15_4310_a951_08b0ca8ef58a.slice/crio-53104f0b6aa1da8cc43c5c486725d41bb77606ac29dff2c800df15b8c2df89cd WatchSource:0}: Error finding container 53104f0b6aa1da8cc43c5c486725d41bb77606ac29dff2c800df15b8c2df89cd: Status 404 returned error can't find the container with id 53104f0b6aa1da8cc43c5c486725d41bb77606ac29dff2c800df15b8c2df89cd Dec 05 12:27:00 crc kubenswrapper[4807]: I1205 12:27:00.967924 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" event={"ID":"a3d95952-eefb-4b02-8480-aa8624398190","Type":"ContainerStarted","Data":"d7268cb060633909a08534442da1e772e28fe19ea6f03724716375bac2a58efc"} Dec 05 12:27:01 crc kubenswrapper[4807]: I1205 12:27:01.743185 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:01 crc kubenswrapper[4807]: I1205 12:27:01.991850 4807 generic.go:334] "Generic (PLEG): container finished" podID="d1bef399-bb15-4310-a951-08b0ca8ef58a" containerID="736c031f3b1359935049cac1b96aca30b54d59751b07a0a4c98167bded987d62" exitCode=0 Dec 05 12:27:01 crc kubenswrapper[4807]: I1205 12:27:01.991908 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-create-22lp6" event={"ID":"d1bef399-bb15-4310-a951-08b0ca8ef58a","Type":"ContainerDied","Data":"736c031f3b1359935049cac1b96aca30b54d59751b07a0a4c98167bded987d62"} Dec 05 12:27:01 crc kubenswrapper[4807]: I1205 12:27:01.992301 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-create-22lp6" event={"ID":"d1bef399-bb15-4310-a951-08b0ca8ef58a","Type":"ContainerStarted","Data":"53104f0b6aa1da8cc43c5c486725d41bb77606ac29dff2c800df15b8c2df89cd"} Dec 05 12:27:01 crc kubenswrapper[4807]: I1205 12:27:01.994641 4807 generic.go:334] "Generic (PLEG): container finished" podID="a3d95952-eefb-4b02-8480-aa8624398190" containerID="01ab24bdff591ff585f628273fab1e40a730323ec47aa8c84c4a5a3d34ad4343" exitCode=0 Dec 05 12:27:01 crc kubenswrapper[4807]: I1205 12:27:01.994685 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" event={"ID":"a3d95952-eefb-4b02-8480-aa8624398190","Type":"ContainerDied","Data":"01ab24bdff591ff585f628273fab1e40a730323ec47aa8c84c4a5a3d34ad4343"} Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.003215 4807 generic.go:334] "Generic (PLEG): container finished" podID="3c5fe773-7fd0-4c59-b0c9-1391e8944c1e" containerID="4455aeb6c7803f06a7d02efc3b40e0744ac038528aab217bcbcaa6d51ade00d4" exitCode=0 Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.003332 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e","Type":"ContainerDied","Data":"4455aeb6c7803f06a7d02efc3b40e0744ac038528aab217bcbcaa6d51ade00d4"} Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.319665 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.324440 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.465659 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87l5p\" (UniqueName: \"kubernetes.io/projected/d1bef399-bb15-4310-a951-08b0ca8ef58a-kube-api-access-87l5p\") pod \"d1bef399-bb15-4310-a951-08b0ca8ef58a\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.465800 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d95952-eefb-4b02-8480-aa8624398190-operator-scripts\") pod \"a3d95952-eefb-4b02-8480-aa8624398190\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.465838 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvz5v\" (UniqueName: \"kubernetes.io/projected/a3d95952-eefb-4b02-8480-aa8624398190-kube-api-access-tvz5v\") pod \"a3d95952-eefb-4b02-8480-aa8624398190\" (UID: \"a3d95952-eefb-4b02-8480-aa8624398190\") " Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.465862 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1bef399-bb15-4310-a951-08b0ca8ef58a-operator-scripts\") pod \"d1bef399-bb15-4310-a951-08b0ca8ef58a\" (UID: \"d1bef399-bb15-4310-a951-08b0ca8ef58a\") " Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.466182 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d95952-eefb-4b02-8480-aa8624398190-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3d95952-eefb-4b02-8480-aa8624398190" (UID: "a3d95952-eefb-4b02-8480-aa8624398190"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.466494 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d95952-eefb-4b02-8480-aa8624398190-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.466644 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1bef399-bb15-4310-a951-08b0ca8ef58a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1bef399-bb15-4310-a951-08b0ca8ef58a" (UID: "d1bef399-bb15-4310-a951-08b0ca8ef58a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.470653 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1bef399-bb15-4310-a951-08b0ca8ef58a-kube-api-access-87l5p" (OuterVolumeSpecName: "kube-api-access-87l5p") pod "d1bef399-bb15-4310-a951-08b0ca8ef58a" (UID: "d1bef399-bb15-4310-a951-08b0ca8ef58a"). InnerVolumeSpecName "kube-api-access-87l5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.470785 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3d95952-eefb-4b02-8480-aa8624398190-kube-api-access-tvz5v" (OuterVolumeSpecName: "kube-api-access-tvz5v") pod "a3d95952-eefb-4b02-8480-aa8624398190" (UID: "a3d95952-eefb-4b02-8480-aa8624398190"). InnerVolumeSpecName "kube-api-access-tvz5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.567885 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvz5v\" (UniqueName: \"kubernetes.io/projected/a3d95952-eefb-4b02-8480-aa8624398190-kube-api-access-tvz5v\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.567941 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1bef399-bb15-4310-a951-08b0ca8ef58a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:03 crc kubenswrapper[4807]: I1205 12:27:03.567954 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87l5p\" (UniqueName: \"kubernetes.io/projected/d1bef399-bb15-4310-a951-08b0ca8ef58a-kube-api-access-87l5p\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.012168 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" event={"ID":"a3d95952-eefb-4b02-8480-aa8624398190","Type":"ContainerDied","Data":"d7268cb060633909a08534442da1e772e28fe19ea6f03724716375bac2a58efc"} Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.012220 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7268cb060633909a08534442da1e772e28fe19ea6f03724716375bac2a58efc" Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.012289 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc" Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.020829 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"3c5fe773-7fd0-4c59-b0c9-1391e8944c1e","Type":"ContainerStarted","Data":"4f616ce7e91d7e12fd5233d5124043ae2c39f1143a439d773dc941ad17d19853"} Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.022046 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.023855 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-22lp6" Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.025016 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-create-22lp6" event={"ID":"d1bef399-bb15-4310-a951-08b0ca8ef58a","Type":"ContainerDied","Data":"53104f0b6aa1da8cc43c5c486725d41bb77606ac29dff2c800df15b8c2df89cd"} Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.025179 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53104f0b6aa1da8cc43c5c486725d41bb77606ac29dff2c800df15b8c2df89cd" Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.026234 4807 generic.go:334] "Generic (PLEG): container finished" podID="74aad9a5-73e5-45a4-9087-c245511adcf0" containerID="a644cbfb8bfd2f1cae8c0c01e4ca0d7a57038f0dc0ae6df6c4571def9d71add8" exitCode=0 Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.026338 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"74aad9a5-73e5-45a4-9087-c245511adcf0","Type":"ContainerDied","Data":"a644cbfb8bfd2f1cae8c0c01e4ca0d7a57038f0dc0ae6df6c4571def9d71add8"} Dec 05 12:27:04 crc kubenswrapper[4807]: I1205 12:27:04.061466 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" podStartSLOduration=38.506397359 podStartE2EDuration="57.061380223s" podCreationTimestamp="2025-12-05 12:26:07 +0000 UTC" firstStartedPulling="2025-12-05 12:26:08.928953951 +0000 UTC m=+1198.422817220" lastFinishedPulling="2025-12-05 12:26:27.483936815 +0000 UTC m=+1216.977800084" observedRunningTime="2025-12-05 12:27:04.057747184 +0000 UTC m=+1253.551610473" watchObservedRunningTime="2025-12-05 12:27:04.061380223 +0000 UTC m=+1253.555243492" Dec 05 12:27:05 crc kubenswrapper[4807]: I1205 12:27:05.038485 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"74aad9a5-73e5-45a4-9087-c245511adcf0","Type":"ContainerStarted","Data":"ce79b88e7e52f96e4696c52f94b4f41280a8e23a8244264d8c5e4057203d0a7e"} Dec 05 12:27:05 crc kubenswrapper[4807]: I1205 12:27:05.040372 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:27:05 crc kubenswrapper[4807]: I1205 12:27:05.068380 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/rabbitmq-server-0" podStartSLOduration=40.413274107 podStartE2EDuration="59.068359473s" podCreationTimestamp="2025-12-05 12:26:06 +0000 UTC" firstStartedPulling="2025-12-05 12:26:09.023119909 +0000 UTC m=+1198.516983178" lastFinishedPulling="2025-12-05 12:26:27.678205275 +0000 UTC m=+1217.172068544" observedRunningTime="2025-12-05 12:27:05.066609121 +0000 UTC m=+1254.560472400" watchObservedRunningTime="2025-12-05 12:27:05.068359473 +0000 UTC m=+1254.562222742" Dec 05 12:27:11 crc kubenswrapper[4807]: I1205 12:27:11.742947 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:11 crc kubenswrapper[4807]: I1205 12:27:11.745491 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:12 crc kubenswrapper[4807]: I1205 12:27:12.094338 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:14 crc kubenswrapper[4807]: I1205 12:27:14.631329 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:27:14 crc kubenswrapper[4807]: I1205 12:27:14.631829 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/prometheus-metric-storage-0" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="prometheus" containerID="cri-o://a49e8a47a02b652f5cfcb766babe1fb648e69e2cf743cc589afa0d61bbcb0c4c" gracePeriod=600 Dec 05 12:27:14 crc kubenswrapper[4807]: I1205 12:27:14.631941 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/prometheus-metric-storage-0" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="config-reloader" containerID="cri-o://d621c6c67e81d7292a9c860401f5dfb4d82982e4ef73bfb1a374f496dc9a29e3" gracePeriod=600 Dec 05 12:27:14 crc kubenswrapper[4807]: I1205 12:27:14.631976 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/prometheus-metric-storage-0" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="thanos-sidecar" containerID="cri-o://37d2a41cbe3f3de81af86995579cc202e4a63bdf6c2a82d79fca6fbea644f650" gracePeriod=600 Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.117935 4807 generic.go:334] "Generic (PLEG): container finished" podID="2774d099-8fed-443a-9e19-5200c7d41f37" containerID="37d2a41cbe3f3de81af86995579cc202e4a63bdf6c2a82d79fca6fbea644f650" exitCode=0 Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.118268 4807 generic.go:334] "Generic (PLEG): container finished" podID="2774d099-8fed-443a-9e19-5200c7d41f37" containerID="d621c6c67e81d7292a9c860401f5dfb4d82982e4ef73bfb1a374f496dc9a29e3" exitCode=0 Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.118281 4807 generic.go:334] "Generic (PLEG): container finished" podID="2774d099-8fed-443a-9e19-5200c7d41f37" containerID="a49e8a47a02b652f5cfcb766babe1fb648e69e2cf743cc589afa0d61bbcb0c4c" exitCode=0 Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.118302 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerDied","Data":"37d2a41cbe3f3de81af86995579cc202e4a63bdf6c2a82d79fca6fbea644f650"} Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.118330 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerDied","Data":"d621c6c67e81d7292a9c860401f5dfb4d82982e4ef73bfb1a374f496dc9a29e3"} Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.118344 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerDied","Data":"a49e8a47a02b652f5cfcb766babe1fb648e69e2cf743cc589afa0d61bbcb0c4c"} Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.689890 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.758737 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2774d099-8fed-443a-9e19-5200c7d41f37-prometheus-metric-storage-rulefiles-0\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.758882 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-tls-assets\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.758932 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-web-config\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.758989 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-config\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.759330 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.759413 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2774d099-8fed-443a-9e19-5200c7d41f37-config-out\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.759541 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-thanos-prometheus-http-client-file\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.759682 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dffg6\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-kube-api-access-dffg6\") pod \"2774d099-8fed-443a-9e19-5200c7d41f37\" (UID: \"2774d099-8fed-443a-9e19-5200c7d41f37\") " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.759945 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2774d099-8fed-443a-9e19-5200c7d41f37-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.761112 4807 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2774d099-8fed-443a-9e19-5200c7d41f37-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.767370 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.769588 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2774d099-8fed-443a-9e19-5200c7d41f37-config-out" (OuterVolumeSpecName: "config-out") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.772889 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.774182 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-kube-api-access-dffg6" (OuterVolumeSpecName: "kube-api-access-dffg6") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "kube-api-access-dffg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.785346 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-config" (OuterVolumeSpecName: "config") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.789609 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.798992 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-web-config" (OuterVolumeSpecName: "web-config") pod "2774d099-8fed-443a-9e19-5200c7d41f37" (UID: "2774d099-8fed-443a-9e19-5200c7d41f37"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.862106 4807 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2774d099-8fed-443a-9e19-5200c7d41f37-config-out\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.862717 4807 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.862817 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dffg6\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-kube-api-access-dffg6\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.862896 4807 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2774d099-8fed-443a-9e19-5200c7d41f37-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.862955 4807 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-web-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.863013 4807 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2774d099-8fed-443a-9e19-5200c7d41f37-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.863101 4807 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") on node \"crc\" " Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.884341 4807 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.884683 4807 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b") on node "crc" Dec 05 12:27:15 crc kubenswrapper[4807]: I1205 12:27:15.965151 4807 reconciler_common.go:293] "Volume detached for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.130380 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"2774d099-8fed-443a-9e19-5200c7d41f37","Type":"ContainerDied","Data":"fdf719874db513c7b9a5c36fe2bbb605aa7cc00a26373972d85a18e03095f3dc"} Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.130438 4807 scope.go:117] "RemoveContainer" containerID="37d2a41cbe3f3de81af86995579cc202e4a63bdf6c2a82d79fca6fbea644f650" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.130603 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.173185 4807 scope.go:117] "RemoveContainer" containerID="d621c6c67e81d7292a9c860401f5dfb4d82982e4ef73bfb1a374f496dc9a29e3" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.185589 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.193018 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.200902 4807 scope.go:117] "RemoveContainer" containerID="a49e8a47a02b652f5cfcb766babe1fb648e69e2cf743cc589afa0d61bbcb0c4c" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.223372 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:27:16 crc kubenswrapper[4807]: E1205 12:27:16.223775 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d95952-eefb-4b02-8480-aa8624398190" containerName="mariadb-account-create-update" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.223794 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d95952-eefb-4b02-8480-aa8624398190" containerName="mariadb-account-create-update" Dec 05 12:27:16 crc kubenswrapper[4807]: E1205 12:27:16.223810 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="thanos-sidecar" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.223819 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="thanos-sidecar" Dec 05 12:27:16 crc kubenswrapper[4807]: E1205 12:27:16.223832 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="prometheus" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.223839 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="prometheus" Dec 05 12:27:16 crc kubenswrapper[4807]: E1205 12:27:16.223854 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1bef399-bb15-4310-a951-08b0ca8ef58a" containerName="mariadb-database-create" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.223863 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1bef399-bb15-4310-a951-08b0ca8ef58a" containerName="mariadb-database-create" Dec 05 12:27:16 crc kubenswrapper[4807]: E1205 12:27:16.223877 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="config-reloader" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.223884 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="config-reloader" Dec 05 12:27:16 crc kubenswrapper[4807]: E1205 12:27:16.223894 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="init-config-reloader" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.223901 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="init-config-reloader" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.224075 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d95952-eefb-4b02-8480-aa8624398190" containerName="mariadb-account-create-update" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.224091 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="thanos-sidecar" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.224110 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="config-reloader" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.224118 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" containerName="prometheus" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.224130 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1bef399-bb15-4310-a951-08b0ca8ef58a" containerName="mariadb-database-create" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.227098 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.228723 4807 scope.go:117] "RemoveContainer" containerID="718faddb357f7c505739b3d88f49e913cd19aac5c9d06ca8e971fcf68125f5c8" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.235162 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.235741 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"metric-storage-prometheus-dockercfg-rwh5f" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.236078 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-web-config" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.236233 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.236450 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.236593 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-metric-storage-prometheus-svc" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.238646 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-tls-assets-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.256412 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269099 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269178 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbnw8\" (UniqueName: \"kubernetes.io/projected/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-kube-api-access-pbnw8\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269217 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269314 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269343 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269390 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269484 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269567 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269602 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269629 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.269654 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-config\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370729 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370793 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbnw8\" (UniqueName: \"kubernetes.io/projected/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-kube-api-access-pbnw8\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370826 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370870 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370894 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370924 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370952 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.370988 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.371017 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.371035 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.371051 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-config\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.372264 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.375590 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.376108 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.376408 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.377046 4807 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.377082 4807 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/76233c9f8195f01138e9549fd668a524184de5a9328019d10824c06f86e8a5fa/globalmount\"" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.377366 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.377950 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.381755 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-config\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.386995 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.390318 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbnw8\" (UniqueName: \"kubernetes.io/projected/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-kube-api-access-pbnw8\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.390903 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4eba60-f9fe-4d26-b6f6-76a326b3e95d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.405762 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5299e1e0-f371-432f-8fb8-975468b4cd8b\") pod \"prometheus-metric-storage-0\" (UID: \"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:16 crc kubenswrapper[4807]: I1205 12:27:16.563747 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:27:17 crc kubenswrapper[4807]: I1205 12:27:17.056221 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Dec 05 12:27:17 crc kubenswrapper[4807]: I1205 12:27:17.143337 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d","Type":"ContainerStarted","Data":"abd5ae75c79bdf07b94d2b5b8bfd54f793c4a48ac4c7a045b225f4a0dbcd1021"} Dec 05 12:27:17 crc kubenswrapper[4807]: I1205 12:27:17.245702 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2774d099-8fed-443a-9e19-5200c7d41f37" path="/var/lib/kubelet/pods/2774d099-8fed-443a-9e19-5200c7d41f37/volumes" Dec 05 12:27:18 crc kubenswrapper[4807]: I1205 12:27:18.436787 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Dec 05 12:27:18 crc kubenswrapper[4807]: I1205 12:27:18.499313 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/rabbitmq-server-0" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.171481 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d","Type":"ContainerStarted","Data":"e6f4e4fd2fe8b820569d2e98de13c52553cfe4d4b81c82aadf4e6d6837f4dadf"} Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.477849 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kmv6v"] Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.478856 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.480823 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.480837 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.481969 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.487100 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-gvjsx" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.492429 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kmv6v"] Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.634697 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbhqj\" (UniqueName: \"kubernetes.io/projected/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-kube-api-access-qbhqj\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.634788 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-config-data\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.634809 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-combined-ca-bundle\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.736041 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbhqj\" (UniqueName: \"kubernetes.io/projected/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-kube-api-access-qbhqj\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.736187 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-config-data\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.736221 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-combined-ca-bundle\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.742890 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-config-data\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.752027 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-combined-ca-bundle\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.752259 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbhqj\" (UniqueName: \"kubernetes.io/projected/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-kube-api-access-qbhqj\") pod \"keystone-db-sync-kmv6v\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:20 crc kubenswrapper[4807]: I1205 12:27:20.795325 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:27:21 crc kubenswrapper[4807]: I1205 12:27:21.281294 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kmv6v"] Dec 05 12:27:22 crc kubenswrapper[4807]: I1205 12:27:22.187418 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" event={"ID":"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6","Type":"ContainerStarted","Data":"295f765ce35997327c9b3c98da791410a43d2c62a84e0d243ab0a12928eb15d0"} Dec 05 12:27:26 crc kubenswrapper[4807]: I1205 12:27:26.229447 4807 generic.go:334] "Generic (PLEG): container finished" podID="6f4eba60-f9fe-4d26-b6f6-76a326b3e95d" containerID="e6f4e4fd2fe8b820569d2e98de13c52553cfe4d4b81c82aadf4e6d6837f4dadf" exitCode=0 Dec 05 12:27:26 crc kubenswrapper[4807]: I1205 12:27:26.229563 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d","Type":"ContainerDied","Data":"e6f4e4fd2fe8b820569d2e98de13c52553cfe4d4b81c82aadf4e6d6837f4dadf"} Dec 05 12:27:27 crc kubenswrapper[4807]: I1205 12:27:27.245227 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d","Type":"ContainerStarted","Data":"40a263377bfc7f864cd1ff5e3b9aed4d8bc5a6784845974ff7d307cca8a61d71"} Dec 05 12:27:43 crc kubenswrapper[4807]: E1205 12:27:43.718128 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Dec 05 12:27:43 crc kubenswrapper[4807]: E1205 12:27:43.718639 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qbhqj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-kmv6v_watcher-kuttl-default(c70e7072-33ae-46e7-9cae-7a7c9b9e80a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 12:27:43 crc kubenswrapper[4807]: E1205 12:27:43.719708 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" podUID="c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" Dec 05 12:27:44 crc kubenswrapper[4807]: E1205 12:27:44.374307 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" podUID="c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" Dec 05 12:27:47 crc kubenswrapper[4807]: I1205 12:27:47.395971 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d","Type":"ContainerStarted","Data":"e3b12bd32c563cddcbe12e2a3bd12ecf39720bdc616ecdd0e5da3f355aba925e"} Dec 05 12:27:47 crc kubenswrapper[4807]: I1205 12:27:47.396604 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"6f4eba60-f9fe-4d26-b6f6-76a326b3e95d","Type":"ContainerStarted","Data":"d5f6290b66b59783cb6321b09a43faa497d3a068bfa0ef732c9190ddec1487c1"} Dec 05 12:27:47 crc kubenswrapper[4807]: I1205 12:27:47.428322 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/prometheus-metric-storage-0" podStartSLOduration=31.428297122 podStartE2EDuration="31.428297122s" podCreationTimestamp="2025-12-05 12:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:27:47.424019867 +0000 UTC m=+1296.917883146" watchObservedRunningTime="2025-12-05 12:27:47.428297122 +0000 UTC m=+1296.922160391" Dec 05 12:27:51 crc kubenswrapper[4807]: I1205 12:27:51.565593 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:28:00 crc kubenswrapper[4807]: I1205 12:28:00.500430 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" event={"ID":"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6","Type":"ContainerStarted","Data":"b3a5525bfa1b8468b7b4468b15fa875ad0647bb165b67b8e04b37df5cec635cf"} Dec 05 12:28:00 crc kubenswrapper[4807]: I1205 12:28:00.520228 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" podStartSLOduration=2.339589359 podStartE2EDuration="40.520206875s" podCreationTimestamp="2025-12-05 12:27:20 +0000 UTC" firstStartedPulling="2025-12-05 12:27:21.292981305 +0000 UTC m=+1270.786844584" lastFinishedPulling="2025-12-05 12:27:59.473598831 +0000 UTC m=+1308.967462100" observedRunningTime="2025-12-05 12:28:00.513590653 +0000 UTC m=+1310.007453942" watchObservedRunningTime="2025-12-05 12:28:00.520206875 +0000 UTC m=+1310.014070144" Dec 05 12:28:01 crc kubenswrapper[4807]: I1205 12:28:01.564671 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:28:01 crc kubenswrapper[4807]: I1205 12:28:01.569899 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:28:02 crc kubenswrapper[4807]: I1205 12:28:02.524286 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/prometheus-metric-storage-0" Dec 05 12:28:05 crc kubenswrapper[4807]: I1205 12:28:05.540743 4807 generic.go:334] "Generic (PLEG): container finished" podID="c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" containerID="b3a5525bfa1b8468b7b4468b15fa875ad0647bb165b67b8e04b37df5cec635cf" exitCode=0 Dec 05 12:28:05 crc kubenswrapper[4807]: I1205 12:28:05.540824 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" event={"ID":"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6","Type":"ContainerDied","Data":"b3a5525bfa1b8468b7b4468b15fa875ad0647bb165b67b8e04b37df5cec635cf"} Dec 05 12:28:06 crc kubenswrapper[4807]: I1205 12:28:06.865644 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.042366 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbhqj\" (UniqueName: \"kubernetes.io/projected/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-kube-api-access-qbhqj\") pod \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.042500 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-combined-ca-bundle\") pod \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.042619 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-config-data\") pod \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\" (UID: \"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6\") " Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.048049 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-kube-api-access-qbhqj" (OuterVolumeSpecName: "kube-api-access-qbhqj") pod "c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" (UID: "c70e7072-33ae-46e7-9cae-7a7c9b9e80a6"). InnerVolumeSpecName "kube-api-access-qbhqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.065199 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" (UID: "c70e7072-33ae-46e7-9cae-7a7c9b9e80a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.089117 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-config-data" (OuterVolumeSpecName: "config-data") pod "c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" (UID: "c70e7072-33ae-46e7-9cae-7a7c9b9e80a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.145063 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbhqj\" (UniqueName: \"kubernetes.io/projected/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-kube-api-access-qbhqj\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.145112 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.145125 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.556797 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" event={"ID":"c70e7072-33ae-46e7-9cae-7a7c9b9e80a6","Type":"ContainerDied","Data":"295f765ce35997327c9b3c98da791410a43d2c62a84e0d243ab0a12928eb15d0"} Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.556834 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="295f765ce35997327c9b3c98da791410a43d2c62a84e0d243ab0a12928eb15d0" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.556846 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kmv6v" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.790324 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-rn64f"] Dec 05 12:28:07 crc kubenswrapper[4807]: E1205 12:28:07.791075 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" containerName="keystone-db-sync" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.791212 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" containerName="keystone-db-sync" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.791562 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" containerName="keystone-db-sync" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.793334 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.803116 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.803200 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"osp-secret" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.803310 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-gvjsx" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.803313 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.803479 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.814511 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-rn64f"] Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.941821 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.944204 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.957922 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.958270 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.959245 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.959359 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-combined-ca-bundle\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.959454 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-credential-keys\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.959593 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w95zk\" (UniqueName: \"kubernetes.io/projected/15dcd857-76ba-42a1-a586-307a5f71dd7f-kube-api-access-w95zk\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.959671 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-fernet-keys\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.959751 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-scripts\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:07 crc kubenswrapper[4807]: I1205 12:28:07.959839 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-config-data\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.061752 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-log-httpd\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.062102 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.062285 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w95zk\" (UniqueName: \"kubernetes.io/projected/15dcd857-76ba-42a1-a586-307a5f71dd7f-kube-api-access-w95zk\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.062706 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-scripts\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.062923 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-scripts\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.063601 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-config-data\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.063760 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-config-data\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.063879 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.063980 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-run-httpd\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.064111 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcpgp\" (UniqueName: \"kubernetes.io/projected/62fb79fd-c507-4bd1-b476-50b734dd941c-kube-api-access-vcpgp\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.064272 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-fernet-keys\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.064428 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-combined-ca-bundle\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.064578 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-credential-keys\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.069072 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-scripts\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.069777 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-credential-keys\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.072062 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-config-data\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.072746 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-fernet-keys\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.087312 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-combined-ca-bundle\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.088602 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w95zk\" (UniqueName: \"kubernetes.io/projected/15dcd857-76ba-42a1-a586-307a5f71dd7f-kube-api-access-w95zk\") pod \"keystone-bootstrap-rn64f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.112955 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.165787 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.165861 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-scripts\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.165905 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-config-data\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.165938 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.165954 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-run-httpd\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.165978 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcpgp\" (UniqueName: \"kubernetes.io/projected/62fb79fd-c507-4bd1-b476-50b734dd941c-kube-api-access-vcpgp\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.166037 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-log-httpd\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.166666 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-log-httpd\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.167140 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-run-httpd\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.171295 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.172672 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.173427 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-config-data\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.182502 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-scripts\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.186884 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcpgp\" (UniqueName: \"kubernetes.io/projected/62fb79fd-c507-4bd1-b476-50b734dd941c-kube-api-access-vcpgp\") pod \"ceilometer-0\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.271356 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.609284 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-rn64f"] Dec 05 12:28:08 crc kubenswrapper[4807]: I1205 12:28:08.823856 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:09 crc kubenswrapper[4807]: I1205 12:28:09.573842 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerStarted","Data":"4ae28281fb0d0c98231a65fe27df1b50f3a1d31454f925ce0196d794626604d6"} Dec 05 12:28:09 crc kubenswrapper[4807]: I1205 12:28:09.575013 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" event={"ID":"15dcd857-76ba-42a1-a586-307a5f71dd7f","Type":"ContainerStarted","Data":"041deaf4a5ec1a298e141f1ecac69a5ae37c79dd2afb79974462ca98bcc879ca"} Dec 05 12:28:09 crc kubenswrapper[4807]: I1205 12:28:09.575033 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" event={"ID":"15dcd857-76ba-42a1-a586-307a5f71dd7f","Type":"ContainerStarted","Data":"3e8abd911daf2de6cf9e626388b8c0b83a82b29a9030eb07e77c3ca448a2daf2"} Dec 05 12:28:09 crc kubenswrapper[4807]: I1205 12:28:09.593064 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" podStartSLOduration=2.593043646 podStartE2EDuration="2.593043646s" podCreationTimestamp="2025-12-05 12:28:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:28:09.592340849 +0000 UTC m=+1319.086204118" watchObservedRunningTime="2025-12-05 12:28:09.593043646 +0000 UTC m=+1319.086906915" Dec 05 12:28:11 crc kubenswrapper[4807]: I1205 12:28:11.335008 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:13 crc kubenswrapper[4807]: I1205 12:28:13.620085 4807 generic.go:334] "Generic (PLEG): container finished" podID="15dcd857-76ba-42a1-a586-307a5f71dd7f" containerID="041deaf4a5ec1a298e141f1ecac69a5ae37c79dd2afb79974462ca98bcc879ca" exitCode=0 Dec 05 12:28:13 crc kubenswrapper[4807]: I1205 12:28:13.620170 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" event={"ID":"15dcd857-76ba-42a1-a586-307a5f71dd7f","Type":"ContainerDied","Data":"041deaf4a5ec1a298e141f1ecac69a5ae37c79dd2afb79974462ca98bcc879ca"} Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.025231 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.193304 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-fernet-keys\") pod \"15dcd857-76ba-42a1-a586-307a5f71dd7f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.193682 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-config-data\") pod \"15dcd857-76ba-42a1-a586-307a5f71dd7f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.193736 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-scripts\") pod \"15dcd857-76ba-42a1-a586-307a5f71dd7f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.193780 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-credential-keys\") pod \"15dcd857-76ba-42a1-a586-307a5f71dd7f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.193817 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-combined-ca-bundle\") pod \"15dcd857-76ba-42a1-a586-307a5f71dd7f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.193854 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w95zk\" (UniqueName: \"kubernetes.io/projected/15dcd857-76ba-42a1-a586-307a5f71dd7f-kube-api-access-w95zk\") pod \"15dcd857-76ba-42a1-a586-307a5f71dd7f\" (UID: \"15dcd857-76ba-42a1-a586-307a5f71dd7f\") " Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.198296 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "15dcd857-76ba-42a1-a586-307a5f71dd7f" (UID: "15dcd857-76ba-42a1-a586-307a5f71dd7f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.198490 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-scripts" (OuterVolumeSpecName: "scripts") pod "15dcd857-76ba-42a1-a586-307a5f71dd7f" (UID: "15dcd857-76ba-42a1-a586-307a5f71dd7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.199720 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15dcd857-76ba-42a1-a586-307a5f71dd7f-kube-api-access-w95zk" (OuterVolumeSpecName: "kube-api-access-w95zk") pod "15dcd857-76ba-42a1-a586-307a5f71dd7f" (UID: "15dcd857-76ba-42a1-a586-307a5f71dd7f"). InnerVolumeSpecName "kube-api-access-w95zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.203793 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "15dcd857-76ba-42a1-a586-307a5f71dd7f" (UID: "15dcd857-76ba-42a1-a586-307a5f71dd7f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.229484 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15dcd857-76ba-42a1-a586-307a5f71dd7f" (UID: "15dcd857-76ba-42a1-a586-307a5f71dd7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.232314 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-config-data" (OuterVolumeSpecName: "config-data") pod "15dcd857-76ba-42a1-a586-307a5f71dd7f" (UID: "15dcd857-76ba-42a1-a586-307a5f71dd7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.296659 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.296734 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.296750 4807 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.296765 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.296777 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w95zk\" (UniqueName: \"kubernetes.io/projected/15dcd857-76ba-42a1-a586-307a5f71dd7f-kube-api-access-w95zk\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.296969 4807 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15dcd857-76ba-42a1-a586-307a5f71dd7f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.657329 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.657324 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-rn64f" event={"ID":"15dcd857-76ba-42a1-a586-307a5f71dd7f","Type":"ContainerDied","Data":"3e8abd911daf2de6cf9e626388b8c0b83a82b29a9030eb07e77c3ca448a2daf2"} Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.657487 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e8abd911daf2de6cf9e626388b8c0b83a82b29a9030eb07e77c3ca448a2daf2" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.666392 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerStarted","Data":"feca82ad69006be0a2e2782907758379b8d91b466c91aac25febdd6595ce74dd"} Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.767314 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-rn64f"] Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.774645 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-rn64f"] Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.839352 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-hznkk"] Dec 05 12:28:15 crc kubenswrapper[4807]: E1205 12:28:15.840014 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15dcd857-76ba-42a1-a586-307a5f71dd7f" containerName="keystone-bootstrap" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.840044 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="15dcd857-76ba-42a1-a586-307a5f71dd7f" containerName="keystone-bootstrap" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.840285 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="15dcd857-76ba-42a1-a586-307a5f71dd7f" containerName="keystone-bootstrap" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.861362 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-hznkk"] Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.861492 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.868183 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.868242 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.868486 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.868585 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-gvjsx" Dec 05 12:28:15 crc kubenswrapper[4807]: I1205 12:28:15.868717 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"osp-secret" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.010649 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-combined-ca-bundle\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.010853 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-fernet-keys\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.010912 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-config-data\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.011008 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsvnx\" (UniqueName: \"kubernetes.io/projected/1b49ba77-5887-4fb4-a553-8b911154f241-kube-api-access-nsvnx\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.011098 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-scripts\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.011275 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-credential-keys\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.112999 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-combined-ca-bundle\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.113072 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-fernet-keys\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.113095 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-config-data\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.113113 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsvnx\" (UniqueName: \"kubernetes.io/projected/1b49ba77-5887-4fb4-a553-8b911154f241-kube-api-access-nsvnx\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.113134 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-scripts\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.113166 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-credential-keys\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.116931 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-scripts\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.117186 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-config-data\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.118735 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-credential-keys\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.122168 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-fernet-keys\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.125845 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-combined-ca-bundle\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.134221 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsvnx\" (UniqueName: \"kubernetes.io/projected/1b49ba77-5887-4fb4-a553-8b911154f241-kube-api-access-nsvnx\") pod \"keystone-bootstrap-hznkk\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:16 crc kubenswrapper[4807]: I1205 12:28:16.183098 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:17 crc kubenswrapper[4807]: I1205 12:28:17.055634 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-hznkk"] Dec 05 12:28:17 crc kubenswrapper[4807]: I1205 12:28:17.246216 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15dcd857-76ba-42a1-a586-307a5f71dd7f" path="/var/lib/kubelet/pods/15dcd857-76ba-42a1-a586-307a5f71dd7f/volumes" Dec 05 12:28:17 crc kubenswrapper[4807]: W1205 12:28:17.361460 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b49ba77_5887_4fb4_a553_8b911154f241.slice/crio-3f0df5a853a4411302fb2304f03892f43eb3a2f38e27ae33fb53b55c5220906e WatchSource:0}: Error finding container 3f0df5a853a4411302fb2304f03892f43eb3a2f38e27ae33fb53b55c5220906e: Status 404 returned error can't find the container with id 3f0df5a853a4411302fb2304f03892f43eb3a2f38e27ae33fb53b55c5220906e Dec 05 12:28:17 crc kubenswrapper[4807]: I1205 12:28:17.681700 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" event={"ID":"1b49ba77-5887-4fb4-a553-8b911154f241","Type":"ContainerStarted","Data":"3f0df5a853a4411302fb2304f03892f43eb3a2f38e27ae33fb53b55c5220906e"} Dec 05 12:28:18 crc kubenswrapper[4807]: I1205 12:28:18.690958 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerStarted","Data":"5093ffd89744eadf6f5d2c19ad617cd1686ad447893e8ba6a3c30f8107df6791"} Dec 05 12:28:18 crc kubenswrapper[4807]: I1205 12:28:18.692505 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" event={"ID":"1b49ba77-5887-4fb4-a553-8b911154f241","Type":"ContainerStarted","Data":"ee15e917dcf67b32d377fa50d07108e9e96742e9f36fe8a806958ddd2a54339d"} Dec 05 12:28:18 crc kubenswrapper[4807]: I1205 12:28:18.711295 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" podStartSLOduration=3.711274892 podStartE2EDuration="3.711274892s" podCreationTimestamp="2025-12-05 12:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:28:18.709127618 +0000 UTC m=+1328.202990887" watchObservedRunningTime="2025-12-05 12:28:18.711274892 +0000 UTC m=+1328.205138161" Dec 05 12:28:23 crc kubenswrapper[4807]: I1205 12:28:23.731112 4807 generic.go:334] "Generic (PLEG): container finished" podID="1b49ba77-5887-4fb4-a553-8b911154f241" containerID="ee15e917dcf67b32d377fa50d07108e9e96742e9f36fe8a806958ddd2a54339d" exitCode=0 Dec 05 12:28:23 crc kubenswrapper[4807]: I1205 12:28:23.731297 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" event={"ID":"1b49ba77-5887-4fb4-a553-8b911154f241","Type":"ContainerDied","Data":"ee15e917dcf67b32d377fa50d07108e9e96742e9f36fe8a806958ddd2a54339d"} Dec 05 12:28:26 crc kubenswrapper[4807]: I1205 12:28:26.992387 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.123013 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-credential-keys\") pod \"1b49ba77-5887-4fb4-a553-8b911154f241\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.123057 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-config-data\") pod \"1b49ba77-5887-4fb4-a553-8b911154f241\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.123083 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsvnx\" (UniqueName: \"kubernetes.io/projected/1b49ba77-5887-4fb4-a553-8b911154f241-kube-api-access-nsvnx\") pod \"1b49ba77-5887-4fb4-a553-8b911154f241\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.123109 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-scripts\") pod \"1b49ba77-5887-4fb4-a553-8b911154f241\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.123150 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-combined-ca-bundle\") pod \"1b49ba77-5887-4fb4-a553-8b911154f241\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.123254 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-fernet-keys\") pod \"1b49ba77-5887-4fb4-a553-8b911154f241\" (UID: \"1b49ba77-5887-4fb4-a553-8b911154f241\") " Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.162438 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-scripts" (OuterVolumeSpecName: "scripts") pod "1b49ba77-5887-4fb4-a553-8b911154f241" (UID: "1b49ba77-5887-4fb4-a553-8b911154f241"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.162600 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1b49ba77-5887-4fb4-a553-8b911154f241" (UID: "1b49ba77-5887-4fb4-a553-8b911154f241"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.162646 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b49ba77-5887-4fb4-a553-8b911154f241-kube-api-access-nsvnx" (OuterVolumeSpecName: "kube-api-access-nsvnx") pod "1b49ba77-5887-4fb4-a553-8b911154f241" (UID: "1b49ba77-5887-4fb4-a553-8b911154f241"). InnerVolumeSpecName "kube-api-access-nsvnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.165690 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1b49ba77-5887-4fb4-a553-8b911154f241" (UID: "1b49ba77-5887-4fb4-a553-8b911154f241"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.184719 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-config-data" (OuterVolumeSpecName: "config-data") pod "1b49ba77-5887-4fb4-a553-8b911154f241" (UID: "1b49ba77-5887-4fb4-a553-8b911154f241"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.216690 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b49ba77-5887-4fb4-a553-8b911154f241" (UID: "1b49ba77-5887-4fb4-a553-8b911154f241"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.227648 4807 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.227677 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.227688 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsvnx\" (UniqueName: \"kubernetes.io/projected/1b49ba77-5887-4fb4-a553-8b911154f241-kube-api-access-nsvnx\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.227698 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.227708 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.227716 4807 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b49ba77-5887-4fb4-a553-8b911154f241-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.763884 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerStarted","Data":"17b544e8bd8dfcc3e05db39629084da4c7131d2cca6e3fccbf4a585a1fed3e6e"} Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.765337 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" event={"ID":"1b49ba77-5887-4fb4-a553-8b911154f241","Type":"ContainerDied","Data":"3f0df5a853a4411302fb2304f03892f43eb3a2f38e27ae33fb53b55c5220906e"} Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.765364 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f0df5a853a4411302fb2304f03892f43eb3a2f38e27ae33fb53b55c5220906e" Dec 05 12:28:27 crc kubenswrapper[4807]: I1205 12:28:27.765406 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-hznkk" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.158208 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-694f96b6d4-c2bm4"] Dec 05 12:28:28 crc kubenswrapper[4807]: E1205 12:28:28.158870 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b49ba77-5887-4fb4-a553-8b911154f241" containerName="keystone-bootstrap" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.158900 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b49ba77-5887-4fb4-a553-8b911154f241" containerName="keystone-bootstrap" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.159203 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b49ba77-5887-4fb4-a553-8b911154f241" containerName="keystone-bootstrap" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.160361 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.163790 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-keystone-public-svc" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.163905 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-keystone-internal-svc" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.164073 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.164768 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.164952 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.165096 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-gvjsx" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.168825 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-694f96b6d4-c2bm4"] Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253320 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-fernet-keys\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253397 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-combined-ca-bundle\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253425 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-config-data\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253471 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-internal-tls-certs\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253487 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-scripts\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253509 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrjkm\" (UniqueName: \"kubernetes.io/projected/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-kube-api-access-lrjkm\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253539 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-public-tls-certs\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.253558 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-credential-keys\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354449 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-combined-ca-bundle\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354503 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-config-data\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354573 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-internal-tls-certs\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354590 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-scripts\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354616 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-public-tls-certs\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354635 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrjkm\" (UniqueName: \"kubernetes.io/projected/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-kube-api-access-lrjkm\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354653 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-credential-keys\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.354697 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-fernet-keys\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.360128 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-scripts\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.360250 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-combined-ca-bundle\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.362409 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-credential-keys\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.362491 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-config-data\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.362506 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-public-tls-certs\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.363414 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-internal-tls-certs\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.365833 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-fernet-keys\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.374284 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrjkm\" (UniqueName: \"kubernetes.io/projected/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-kube-api-access-lrjkm\") pod \"keystone-694f96b6d4-c2bm4\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:28 crc kubenswrapper[4807]: I1205 12:28:28.485870 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:29 crc kubenswrapper[4807]: I1205 12:28:29.010643 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-694f96b6d4-c2bm4"] Dec 05 12:28:29 crc kubenswrapper[4807]: I1205 12:28:29.798877 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" event={"ID":"50f5f9c8-ef02-49d3-9b04-115ee3f34e93","Type":"ContainerStarted","Data":"b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016"} Dec 05 12:28:29 crc kubenswrapper[4807]: I1205 12:28:29.799710 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" event={"ID":"50f5f9c8-ef02-49d3-9b04-115ee3f34e93","Type":"ContainerStarted","Data":"6310e9c6db01fcb9092f9b320d76eb7e07b176147418b2077b0b9dd4d9e7832d"} Dec 05 12:28:29 crc kubenswrapper[4807]: I1205 12:28:29.799731 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:28:29 crc kubenswrapper[4807]: I1205 12:28:29.818602 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" podStartSLOduration=1.818585856 podStartE2EDuration="1.818585856s" podCreationTimestamp="2025-12-05 12:28:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:28:29.815967411 +0000 UTC m=+1339.309830670" watchObservedRunningTime="2025-12-05 12:28:29.818585856 +0000 UTC m=+1339.312449125" Dec 05 12:28:39 crc kubenswrapper[4807]: I1205 12:28:39.907788 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerStarted","Data":"fb45bbdfa4ae03303facda1912ab533e80230753d9e1741f5f343a970044cc41"} Dec 05 12:28:39 crc kubenswrapper[4807]: I1205 12:28:39.908304 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:39 crc kubenswrapper[4807]: I1205 12:28:39.907976 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="sg-core" containerID="cri-o://17b544e8bd8dfcc3e05db39629084da4c7131d2cca6e3fccbf4a585a1fed3e6e" gracePeriod=30 Dec 05 12:28:39 crc kubenswrapper[4807]: I1205 12:28:39.907901 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-central-agent" containerID="cri-o://feca82ad69006be0a2e2782907758379b8d91b466c91aac25febdd6595ce74dd" gracePeriod=30 Dec 05 12:28:39 crc kubenswrapper[4807]: I1205 12:28:39.908090 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-notification-agent" containerID="cri-o://5093ffd89744eadf6f5d2c19ad617cd1686ad447893e8ba6a3c30f8107df6791" gracePeriod=30 Dec 05 12:28:39 crc kubenswrapper[4807]: I1205 12:28:39.908014 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="proxy-httpd" containerID="cri-o://fb45bbdfa4ae03303facda1912ab533e80230753d9e1741f5f343a970044cc41" gracePeriod=30 Dec 05 12:28:39 crc kubenswrapper[4807]: I1205 12:28:39.936751 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.7813561030000002 podStartE2EDuration="32.936730277s" podCreationTimestamp="2025-12-05 12:28:07 +0000 UTC" firstStartedPulling="2025-12-05 12:28:08.83612451 +0000 UTC m=+1318.329987779" lastFinishedPulling="2025-12-05 12:28:38.991498684 +0000 UTC m=+1348.485361953" observedRunningTime="2025-12-05 12:28:39.930937155 +0000 UTC m=+1349.424800434" watchObservedRunningTime="2025-12-05 12:28:39.936730277 +0000 UTC m=+1349.430593546" Dec 05 12:28:40 crc kubenswrapper[4807]: I1205 12:28:40.923048 4807 generic.go:334] "Generic (PLEG): container finished" podID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerID="fb45bbdfa4ae03303facda1912ab533e80230753d9e1741f5f343a970044cc41" exitCode=0 Dec 05 12:28:40 crc kubenswrapper[4807]: I1205 12:28:40.923593 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerDied","Data":"fb45bbdfa4ae03303facda1912ab533e80230753d9e1741f5f343a970044cc41"} Dec 05 12:28:40 crc kubenswrapper[4807]: I1205 12:28:40.923671 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerDied","Data":"17b544e8bd8dfcc3e05db39629084da4c7131d2cca6e3fccbf4a585a1fed3e6e"} Dec 05 12:28:40 crc kubenswrapper[4807]: I1205 12:28:40.923625 4807 generic.go:334] "Generic (PLEG): container finished" podID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerID="17b544e8bd8dfcc3e05db39629084da4c7131d2cca6e3fccbf4a585a1fed3e6e" exitCode=2 Dec 05 12:28:40 crc kubenswrapper[4807]: I1205 12:28:40.923707 4807 generic.go:334] "Generic (PLEG): container finished" podID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerID="feca82ad69006be0a2e2782907758379b8d91b466c91aac25febdd6595ce74dd" exitCode=0 Dec 05 12:28:40 crc kubenswrapper[4807]: I1205 12:28:40.923732 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerDied","Data":"feca82ad69006be0a2e2782907758379b8d91b466c91aac25febdd6595ce74dd"} Dec 05 12:28:41 crc kubenswrapper[4807]: I1205 12:28:41.933764 4807 generic.go:334] "Generic (PLEG): container finished" podID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerID="5093ffd89744eadf6f5d2c19ad617cd1686ad447893e8ba6a3c30f8107df6791" exitCode=0 Dec 05 12:28:41 crc kubenswrapper[4807]: I1205 12:28:41.933951 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerDied","Data":"5093ffd89744eadf6f5d2c19ad617cd1686ad447893e8ba6a3c30f8107df6791"} Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.583756 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.705762 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-run-httpd\") pod \"62fb79fd-c507-4bd1-b476-50b734dd941c\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.705872 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcpgp\" (UniqueName: \"kubernetes.io/projected/62fb79fd-c507-4bd1-b476-50b734dd941c-kube-api-access-vcpgp\") pod \"62fb79fd-c507-4bd1-b476-50b734dd941c\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.705977 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-combined-ca-bundle\") pod \"62fb79fd-c507-4bd1-b476-50b734dd941c\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.706012 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-log-httpd\") pod \"62fb79fd-c507-4bd1-b476-50b734dd941c\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.706038 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-scripts\") pod \"62fb79fd-c507-4bd1-b476-50b734dd941c\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.706069 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-sg-core-conf-yaml\") pod \"62fb79fd-c507-4bd1-b476-50b734dd941c\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.706098 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-config-data\") pod \"62fb79fd-c507-4bd1-b476-50b734dd941c\" (UID: \"62fb79fd-c507-4bd1-b476-50b734dd941c\") " Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.706674 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "62fb79fd-c507-4bd1-b476-50b734dd941c" (UID: "62fb79fd-c507-4bd1-b476-50b734dd941c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.706971 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "62fb79fd-c507-4bd1-b476-50b734dd941c" (UID: "62fb79fd-c507-4bd1-b476-50b734dd941c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.711899 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fb79fd-c507-4bd1-b476-50b734dd941c-kube-api-access-vcpgp" (OuterVolumeSpecName: "kube-api-access-vcpgp") pod "62fb79fd-c507-4bd1-b476-50b734dd941c" (UID: "62fb79fd-c507-4bd1-b476-50b734dd941c"). InnerVolumeSpecName "kube-api-access-vcpgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.712128 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-scripts" (OuterVolumeSpecName: "scripts") pod "62fb79fd-c507-4bd1-b476-50b734dd941c" (UID: "62fb79fd-c507-4bd1-b476-50b734dd941c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.734694 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "62fb79fd-c507-4bd1-b476-50b734dd941c" (UID: "62fb79fd-c507-4bd1-b476-50b734dd941c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.780702 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62fb79fd-c507-4bd1-b476-50b734dd941c" (UID: "62fb79fd-c507-4bd1-b476-50b734dd941c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.807506 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcpgp\" (UniqueName: \"kubernetes.io/projected/62fb79fd-c507-4bd1-b476-50b734dd941c-kube-api-access-vcpgp\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.807557 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.807590 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.807606 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.807623 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.807634 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fb79fd-c507-4bd1-b476-50b734dd941c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.808063 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-config-data" (OuterVolumeSpecName: "config-data") pod "62fb79fd-c507-4bd1-b476-50b734dd941c" (UID: "62fb79fd-c507-4bd1-b476-50b734dd941c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.910299 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fb79fd-c507-4bd1-b476-50b734dd941c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.946038 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"62fb79fd-c507-4bd1-b476-50b734dd941c","Type":"ContainerDied","Data":"4ae28281fb0d0c98231a65fe27df1b50f3a1d31454f925ce0196d794626604d6"} Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.946100 4807 scope.go:117] "RemoveContainer" containerID="fb45bbdfa4ae03303facda1912ab533e80230753d9e1741f5f343a970044cc41" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.946116 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.981063 4807 scope.go:117] "RemoveContainer" containerID="17b544e8bd8dfcc3e05db39629084da4c7131d2cca6e3fccbf4a585a1fed3e6e" Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.982966 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.991472 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:42 crc kubenswrapper[4807]: I1205 12:28:42.999354 4807 scope.go:117] "RemoveContainer" containerID="5093ffd89744eadf6f5d2c19ad617cd1686ad447893e8ba6a3c30f8107df6791" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013118 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:43 crc kubenswrapper[4807]: E1205 12:28:43.013451 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-central-agent" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013469 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-central-agent" Dec 05 12:28:43 crc kubenswrapper[4807]: E1205 12:28:43.013488 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="proxy-httpd" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013496 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="proxy-httpd" Dec 05 12:28:43 crc kubenswrapper[4807]: E1205 12:28:43.013517 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="sg-core" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013530 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="sg-core" Dec 05 12:28:43 crc kubenswrapper[4807]: E1205 12:28:43.013544 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-notification-agent" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013552 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-notification-agent" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013869 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="proxy-httpd" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013899 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-notification-agent" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013921 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="sg-core" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.013930 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" containerName="ceilometer-central-agent" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.015305 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.018700 4807 scope.go:117] "RemoveContainer" containerID="feca82ad69006be0a2e2782907758379b8d91b466c91aac25febdd6595ce74dd" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.019323 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.026590 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.026781 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.113448 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.113500 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-scripts\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.113545 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-config-data\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.113652 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flctw\" (UniqueName: \"kubernetes.io/projected/98c406d3-6570-4956-ad01-fcf83a35f565-kube-api-access-flctw\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.113699 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.113728 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-log-httpd\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.113753 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-run-httpd\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.214876 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.214938 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-scripts\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.214994 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-config-data\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.215050 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flctw\" (UniqueName: \"kubernetes.io/projected/98c406d3-6570-4956-ad01-fcf83a35f565-kube-api-access-flctw\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.215147 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.215181 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-log-httpd\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.215210 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-run-httpd\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.215804 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-run-httpd\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.215828 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-log-httpd\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.218976 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.220265 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-config-data\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.223353 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.232251 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-scripts\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.234802 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flctw\" (UniqueName: \"kubernetes.io/projected/98c406d3-6570-4956-ad01-fcf83a35f565-kube-api-access-flctw\") pod \"ceilometer-0\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.250243 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62fb79fd-c507-4bd1-b476-50b734dd941c" path="/var/lib/kubelet/pods/62fb79fd-c507-4bd1-b476-50b734dd941c/volumes" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.329801 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.793531 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:28:43 crc kubenswrapper[4807]: W1205 12:28:43.804331 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98c406d3_6570_4956_ad01_fcf83a35f565.slice/crio-92a88f677a1d649e63304c0f20d31dee4a0d6f25f4afe7bc608bcfe245143015 WatchSource:0}: Error finding container 92a88f677a1d649e63304c0f20d31dee4a0d6f25f4afe7bc608bcfe245143015: Status 404 returned error can't find the container with id 92a88f677a1d649e63304c0f20d31dee4a0d6f25f4afe7bc608bcfe245143015 Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.808656 4807 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:28:43 crc kubenswrapper[4807]: I1205 12:28:43.956886 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerStarted","Data":"92a88f677a1d649e63304c0f20d31dee4a0d6f25f4afe7bc608bcfe245143015"} Dec 05 12:28:44 crc kubenswrapper[4807]: I1205 12:28:44.971969 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerStarted","Data":"a8e4a3acd7ca4b1a2ef63b40f11fe8b0b61ab3716db3a19f60d674aac9289e71"} Dec 05 12:28:45 crc kubenswrapper[4807]: I1205 12:28:45.984266 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerStarted","Data":"718476f3d7f59eab77a9aff9e3005dd34a8fc033c6e2af044c306b1afb284dd4"} Dec 05 12:28:45 crc kubenswrapper[4807]: I1205 12:28:45.984837 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerStarted","Data":"db6d3c9f2bdab5fcc8f703f696b1ee4f15e91c695eec6cfd6b2a82f26ebdfeed"} Dec 05 12:28:49 crc kubenswrapper[4807]: I1205 12:28:49.017961 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerStarted","Data":"41d09ee567af9fcb05bed6fc1765268c0ee83a349e2ac3626e1d8d2c722f6f5d"} Dec 05 12:28:49 crc kubenswrapper[4807]: I1205 12:28:49.018624 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:28:49 crc kubenswrapper[4807]: I1205 12:28:49.042328 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.163993606 podStartE2EDuration="7.042295261s" podCreationTimestamp="2025-12-05 12:28:42 +0000 UTC" firstStartedPulling="2025-12-05 12:28:43.808353108 +0000 UTC m=+1353.302216387" lastFinishedPulling="2025-12-05 12:28:48.686654753 +0000 UTC m=+1358.180518042" observedRunningTime="2025-12-05 12:28:49.038353294 +0000 UTC m=+1358.532216553" watchObservedRunningTime="2025-12-05 12:28:49.042295261 +0000 UTC m=+1358.536158530" Dec 05 12:28:52 crc kubenswrapper[4807]: I1205 12:28:52.466576 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:28:52 crc kubenswrapper[4807]: I1205 12:28:52.467803 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:29:00 crc kubenswrapper[4807]: I1205 12:29:00.158994 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.668967 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/openstackclient"] Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.670801 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.674358 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openstack-config" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.685523 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"openstack-config-secret" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.686581 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"openstackclient-openstackclient-dockercfg-wqp42" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.699464 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstackclient"] Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.769465 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/118370dc-bda4-46d4-a447-c3022ae95767-openstack-config\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.871169 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/118370dc-bda4-46d4-a447-c3022ae95767-openstack-config\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.871239 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/118370dc-bda4-46d4-a447-c3022ae95767-openstack-config-secret\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.871265 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/118370dc-bda4-46d4-a447-c3022ae95767-combined-ca-bundle\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.871328 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc44p\" (UniqueName: \"kubernetes.io/projected/118370dc-bda4-46d4-a447-c3022ae95767-kube-api-access-gc44p\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.872101 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/118370dc-bda4-46d4-a447-c3022ae95767-openstack-config\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.972919 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc44p\" (UniqueName: \"kubernetes.io/projected/118370dc-bda4-46d4-a447-c3022ae95767-kube-api-access-gc44p\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.973080 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/118370dc-bda4-46d4-a447-c3022ae95767-openstack-config-secret\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.973112 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/118370dc-bda4-46d4-a447-c3022ae95767-combined-ca-bundle\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.979854 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/118370dc-bda4-46d4-a447-c3022ae95767-openstack-config-secret\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.980161 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/118370dc-bda4-46d4-a447-c3022ae95767-combined-ca-bundle\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:04 crc kubenswrapper[4807]: I1205 12:29:04.995312 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc44p\" (UniqueName: \"kubernetes.io/projected/118370dc-bda4-46d4-a447-c3022ae95767-kube-api-access-gc44p\") pod \"openstackclient\" (UID: \"118370dc-bda4-46d4-a447-c3022ae95767\") " pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:05 crc kubenswrapper[4807]: I1205 12:29:05.290800 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Dec 05 12:29:05 crc kubenswrapper[4807]: I1205 12:29:05.751613 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstackclient"] Dec 05 12:29:06 crc kubenswrapper[4807]: I1205 12:29:06.145310 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstackclient" event={"ID":"118370dc-bda4-46d4-a447-c3022ae95767","Type":"ContainerStarted","Data":"fe6b4a3c81dc4fa261866361faf8a572996927d4978db1cea44519cce66d6d37"} Dec 05 12:29:13 crc kubenswrapper[4807]: I1205 12:29:13.334873 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:15 crc kubenswrapper[4807]: I1205 12:29:15.229671 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstackclient" event={"ID":"118370dc-bda4-46d4-a447-c3022ae95767","Type":"ContainerStarted","Data":"6502e51b97b27177f0a6b2b22ac2922f358a66436a660dc4af5c9cd477db1007"} Dec 05 12:29:15 crc kubenswrapper[4807]: I1205 12:29:15.254627 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/openstackclient" podStartSLOduration=2.994316821 podStartE2EDuration="11.254602662s" podCreationTimestamp="2025-12-05 12:29:04 +0000 UTC" firstStartedPulling="2025-12-05 12:29:05.75926789 +0000 UTC m=+1375.253131159" lastFinishedPulling="2025-12-05 12:29:14.019553731 +0000 UTC m=+1383.513417000" observedRunningTime="2025-12-05 12:29:15.251215138 +0000 UTC m=+1384.745078407" watchObservedRunningTime="2025-12-05 12:29:15.254602662 +0000 UTC m=+1384.748465931" Dec 05 12:29:15 crc kubenswrapper[4807]: I1205 12:29:15.809587 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:29:15 crc kubenswrapper[4807]: I1205 12:29:15.810069 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/kube-state-metrics-0" podUID="db9e65be-19f5-4180-9b0c-96f323e547a3" containerName="kube-state-metrics" containerID="cri-o://83172aa6171e6fc96bd9242f36ccaba8d837e325626ba800ca11880c4f62ca35" gracePeriod=30 Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.242333 4807 generic.go:334] "Generic (PLEG): container finished" podID="db9e65be-19f5-4180-9b0c-96f323e547a3" containerID="83172aa6171e6fc96bd9242f36ccaba8d837e325626ba800ca11880c4f62ca35" exitCode=2 Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.242440 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"db9e65be-19f5-4180-9b0c-96f323e547a3","Type":"ContainerDied","Data":"83172aa6171e6fc96bd9242f36ccaba8d837e325626ba800ca11880c4f62ca35"} Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.326930 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.483748 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx76q\" (UniqueName: \"kubernetes.io/projected/db9e65be-19f5-4180-9b0c-96f323e547a3-kube-api-access-sx76q\") pod \"db9e65be-19f5-4180-9b0c-96f323e547a3\" (UID: \"db9e65be-19f5-4180-9b0c-96f323e547a3\") " Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.505320 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db9e65be-19f5-4180-9b0c-96f323e547a3-kube-api-access-sx76q" (OuterVolumeSpecName: "kube-api-access-sx76q") pod "db9e65be-19f5-4180-9b0c-96f323e547a3" (UID: "db9e65be-19f5-4180-9b0c-96f323e547a3"). InnerVolumeSpecName "kube-api-access-sx76q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.586109 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx76q\" (UniqueName: \"kubernetes.io/projected/db9e65be-19f5-4180-9b0c-96f323e547a3-kube-api-access-sx76q\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.814124 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.814800 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-central-agent" containerID="cri-o://a8e4a3acd7ca4b1a2ef63b40f11fe8b0b61ab3716db3a19f60d674aac9289e71" gracePeriod=30 Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.814860 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="sg-core" containerID="cri-o://718476f3d7f59eab77a9aff9e3005dd34a8fc033c6e2af044c306b1afb284dd4" gracePeriod=30 Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.814881 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-notification-agent" containerID="cri-o://db6d3c9f2bdab5fcc8f703f696b1ee4f15e91c695eec6cfd6b2a82f26ebdfeed" gracePeriod=30 Dec 05 12:29:16 crc kubenswrapper[4807]: I1205 12:29:16.814860 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="proxy-httpd" containerID="cri-o://41d09ee567af9fcb05bed6fc1765268c0ee83a349e2ac3626e1d8d2c722f6f5d" gracePeriod=30 Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.250930 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"db9e65be-19f5-4180-9b0c-96f323e547a3","Type":"ContainerDied","Data":"a0be75746c5b0aae5234f80c2ed309682ee0d3f3cf3293ea9c4564b26eec81d2"} Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.250962 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.251009 4807 scope.go:117] "RemoveContainer" containerID="83172aa6171e6fc96bd9242f36ccaba8d837e325626ba800ca11880c4f62ca35" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.255376 4807 generic.go:334] "Generic (PLEG): container finished" podID="98c406d3-6570-4956-ad01-fcf83a35f565" containerID="41d09ee567af9fcb05bed6fc1765268c0ee83a349e2ac3626e1d8d2c722f6f5d" exitCode=0 Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.255408 4807 generic.go:334] "Generic (PLEG): container finished" podID="98c406d3-6570-4956-ad01-fcf83a35f565" containerID="718476f3d7f59eab77a9aff9e3005dd34a8fc033c6e2af044c306b1afb284dd4" exitCode=2 Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.255417 4807 generic.go:334] "Generic (PLEG): container finished" podID="98c406d3-6570-4956-ad01-fcf83a35f565" containerID="a8e4a3acd7ca4b1a2ef63b40f11fe8b0b61ab3716db3a19f60d674aac9289e71" exitCode=0 Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.255436 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerDied","Data":"41d09ee567af9fcb05bed6fc1765268c0ee83a349e2ac3626e1d8d2c722f6f5d"} Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.255460 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerDied","Data":"718476f3d7f59eab77a9aff9e3005dd34a8fc033c6e2af044c306b1afb284dd4"} Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.255470 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerDied","Data":"a8e4a3acd7ca4b1a2ef63b40f11fe8b0b61ab3716db3a19f60d674aac9289e71"} Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.282539 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.289106 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.309834 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:29:17 crc kubenswrapper[4807]: E1205 12:29:17.310424 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9e65be-19f5-4180-9b0c-96f323e547a3" containerName="kube-state-metrics" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.310449 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9e65be-19f5-4180-9b0c-96f323e547a3" containerName="kube-state-metrics" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.310684 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="db9e65be-19f5-4180-9b0c-96f323e547a3" containerName="kube-state-metrics" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.311859 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.314760 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-kube-state-metrics-svc" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.315222 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"kube-state-metrics-tls-config" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.320742 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.403999 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.404060 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.404106 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.404169 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfljn\" (UniqueName: \"kubernetes.io/projected/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-api-access-qfljn\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.505907 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.505974 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.506037 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfljn\" (UniqueName: \"kubernetes.io/projected/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-api-access-qfljn\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.506090 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.515396 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.519332 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.524127 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.537843 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfljn\" (UniqueName: \"kubernetes.io/projected/503ab53a-ec37-4100-9021-a3e7836da8ae-kube-api-access-qfljn\") pod \"kube-state-metrics-0\" (UID: \"503ab53a-ec37-4100-9021-a3e7836da8ae\") " pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:17 crc kubenswrapper[4807]: I1205 12:29:17.638770 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:18 crc kubenswrapper[4807]: I1205 12:29:18.085681 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Dec 05 12:29:18 crc kubenswrapper[4807]: I1205 12:29:18.273257 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"503ab53a-ec37-4100-9021-a3e7836da8ae","Type":"ContainerStarted","Data":"952e97fb5e195e11fbadd78486256f5f512643f197ee6cf650a3d67d47cfa724"} Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.247489 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db9e65be-19f5-4180-9b0c-96f323e547a3" path="/var/lib/kubelet/pods/db9e65be-19f5-4180-9b0c-96f323e547a3/volumes" Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.856423 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r"] Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.857889 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.860005 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.865908 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-5q5fh"] Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.867189 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.875032 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r"] Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.884391 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5q5fh"] Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.947596 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-operator-scripts\") pod \"watcher-db-create-5q5fh\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.947647 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhlwd\" (UniqueName: \"kubernetes.io/projected/3381f31f-572a-4e4f-aae5-d38d9a9253cd-kube-api-access-hhlwd\") pod \"watcher-2ca3-account-create-update-zzd7r\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.947715 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qslcf\" (UniqueName: \"kubernetes.io/projected/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-kube-api-access-qslcf\") pod \"watcher-db-create-5q5fh\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:19 crc kubenswrapper[4807]: I1205 12:29:19.947740 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3381f31f-572a-4e4f-aae5-d38d9a9253cd-operator-scripts\") pod \"watcher-2ca3-account-create-update-zzd7r\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.050506 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-operator-scripts\") pod \"watcher-db-create-5q5fh\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.051204 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhlwd\" (UniqueName: \"kubernetes.io/projected/3381f31f-572a-4e4f-aae5-d38d9a9253cd-kube-api-access-hhlwd\") pod \"watcher-2ca3-account-create-update-zzd7r\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.051323 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-operator-scripts\") pod \"watcher-db-create-5q5fh\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.051665 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qslcf\" (UniqueName: \"kubernetes.io/projected/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-kube-api-access-qslcf\") pod \"watcher-db-create-5q5fh\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.051859 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3381f31f-572a-4e4f-aae5-d38d9a9253cd-operator-scripts\") pod \"watcher-2ca3-account-create-update-zzd7r\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.052494 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3381f31f-572a-4e4f-aae5-d38d9a9253cd-operator-scripts\") pod \"watcher-2ca3-account-create-update-zzd7r\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.073587 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhlwd\" (UniqueName: \"kubernetes.io/projected/3381f31f-572a-4e4f-aae5-d38d9a9253cd-kube-api-access-hhlwd\") pod \"watcher-2ca3-account-create-update-zzd7r\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.083844 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qslcf\" (UniqueName: \"kubernetes.io/projected/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-kube-api-access-qslcf\") pod \"watcher-db-create-5q5fh\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.185642 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.191915 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.642351 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r"] Dec 05 12:29:20 crc kubenswrapper[4807]: I1205 12:29:20.707867 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5q5fh"] Dec 05 12:29:20 crc kubenswrapper[4807]: W1205 12:29:20.709725 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3d9bd6b_3038_4cc3_b204_96b047cfe3ec.slice/crio-9423fb5fcb81b424653f1ee76c50981624a28272f38fd26cf21fd29a08416182 WatchSource:0}: Error finding container 9423fb5fcb81b424653f1ee76c50981624a28272f38fd26cf21fd29a08416182: Status 404 returned error can't find the container with id 9423fb5fcb81b424653f1ee76c50981624a28272f38fd26cf21fd29a08416182 Dec 05 12:29:21 crc kubenswrapper[4807]: I1205 12:29:21.299571 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" event={"ID":"3381f31f-572a-4e4f-aae5-d38d9a9253cd","Type":"ContainerStarted","Data":"52672ff8de092cfdaf33d78ee879de1cc9b55be2d15e741ff3069ce45c0ec094"} Dec 05 12:29:21 crc kubenswrapper[4807]: I1205 12:29:21.301422 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-5q5fh" event={"ID":"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec","Type":"ContainerStarted","Data":"9423fb5fcb81b424653f1ee76c50981624a28272f38fd26cf21fd29a08416182"} Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.313284 4807 generic.go:334] "Generic (PLEG): container finished" podID="98c406d3-6570-4956-ad01-fcf83a35f565" containerID="db6d3c9f2bdab5fcc8f703f696b1ee4f15e91c695eec6cfd6b2a82f26ebdfeed" exitCode=0 Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.313350 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerDied","Data":"db6d3c9f2bdab5fcc8f703f696b1ee4f15e91c695eec6cfd6b2a82f26ebdfeed"} Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.315517 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-5q5fh" event={"ID":"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec","Type":"ContainerStarted","Data":"1e548a724927c630d01a9d517cca27e1992007825cd3f82099c91aa6a04323aa"} Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.318154 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" event={"ID":"3381f31f-572a-4e4f-aae5-d38d9a9253cd","Type":"ContainerStarted","Data":"898fbca429efeb2b2180ddce3fa3cedbb54b0d9ed16e5f37f99f1da44cc99e29"} Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.337630 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-db-create-5q5fh" podStartSLOduration=3.33761367 podStartE2EDuration="3.33761367s" podCreationTimestamp="2025-12-05 12:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:29:22.331450009 +0000 UTC m=+1391.825313278" watchObservedRunningTime="2025-12-05 12:29:22.33761367 +0000 UTC m=+1391.831476929" Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.353742 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" podStartSLOduration=3.353724062 podStartE2EDuration="3.353724062s" podCreationTimestamp="2025-12-05 12:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:29:22.350661127 +0000 UTC m=+1391.844524406" watchObservedRunningTime="2025-12-05 12:29:22.353724062 +0000 UTC m=+1391.847587331" Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.466433 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:29:22 crc kubenswrapper[4807]: I1205 12:29:22.466495 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.056762 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.225573 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flctw\" (UniqueName: \"kubernetes.io/projected/98c406d3-6570-4956-ad01-fcf83a35f565-kube-api-access-flctw\") pod \"98c406d3-6570-4956-ad01-fcf83a35f565\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.225647 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-scripts\") pod \"98c406d3-6570-4956-ad01-fcf83a35f565\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.225799 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-log-httpd\") pod \"98c406d3-6570-4956-ad01-fcf83a35f565\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.225844 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-config-data\") pod \"98c406d3-6570-4956-ad01-fcf83a35f565\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.225934 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-run-httpd\") pod \"98c406d3-6570-4956-ad01-fcf83a35f565\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.225995 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-combined-ca-bundle\") pod \"98c406d3-6570-4956-ad01-fcf83a35f565\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.226039 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-sg-core-conf-yaml\") pod \"98c406d3-6570-4956-ad01-fcf83a35f565\" (UID: \"98c406d3-6570-4956-ad01-fcf83a35f565\") " Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.227817 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "98c406d3-6570-4956-ad01-fcf83a35f565" (UID: "98c406d3-6570-4956-ad01-fcf83a35f565"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.228049 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "98c406d3-6570-4956-ad01-fcf83a35f565" (UID: "98c406d3-6570-4956-ad01-fcf83a35f565"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.231210 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c406d3-6570-4956-ad01-fcf83a35f565-kube-api-access-flctw" (OuterVolumeSpecName: "kube-api-access-flctw") pod "98c406d3-6570-4956-ad01-fcf83a35f565" (UID: "98c406d3-6570-4956-ad01-fcf83a35f565"). InnerVolumeSpecName "kube-api-access-flctw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.231618 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-scripts" (OuterVolumeSpecName: "scripts") pod "98c406d3-6570-4956-ad01-fcf83a35f565" (UID: "98c406d3-6570-4956-ad01-fcf83a35f565"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.253879 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "98c406d3-6570-4956-ad01-fcf83a35f565" (UID: "98c406d3-6570-4956-ad01-fcf83a35f565"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.293504 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98c406d3-6570-4956-ad01-fcf83a35f565" (UID: "98c406d3-6570-4956-ad01-fcf83a35f565"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.308719 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-config-data" (OuterVolumeSpecName: "config-data") pod "98c406d3-6570-4956-ad01-fcf83a35f565" (UID: "98c406d3-6570-4956-ad01-fcf83a35f565"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.338167 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.338209 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.338221 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.338233 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.338244 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flctw\" (UniqueName: \"kubernetes.io/projected/98c406d3-6570-4956-ad01-fcf83a35f565-kube-api-access-flctw\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.338258 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c406d3-6570-4956-ad01-fcf83a35f565-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.338268 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c406d3-6570-4956-ad01-fcf83a35f565-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.343369 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"98c406d3-6570-4956-ad01-fcf83a35f565","Type":"ContainerDied","Data":"92a88f677a1d649e63304c0f20d31dee4a0d6f25f4afe7bc608bcfe245143015"} Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.343432 4807 scope.go:117] "RemoveContainer" containerID="41d09ee567af9fcb05bed6fc1765268c0ee83a349e2ac3626e1d8d2c722f6f5d" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.344121 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.345344 4807 generic.go:334] "Generic (PLEG): container finished" podID="a3d9bd6b-3038-4cc3-b204-96b047cfe3ec" containerID="1e548a724927c630d01a9d517cca27e1992007825cd3f82099c91aa6a04323aa" exitCode=0 Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.345401 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-5q5fh" event={"ID":"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec","Type":"ContainerDied","Data":"1e548a724927c630d01a9d517cca27e1992007825cd3f82099c91aa6a04323aa"} Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.346840 4807 generic.go:334] "Generic (PLEG): container finished" podID="3381f31f-572a-4e4f-aae5-d38d9a9253cd" containerID="898fbca429efeb2b2180ddce3fa3cedbb54b0d9ed16e5f37f99f1da44cc99e29" exitCode=0 Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.346883 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" event={"ID":"3381f31f-572a-4e4f-aae5-d38d9a9253cd","Type":"ContainerDied","Data":"898fbca429efeb2b2180ddce3fa3cedbb54b0d9ed16e5f37f99f1da44cc99e29"} Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.349344 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"503ab53a-ec37-4100-9021-a3e7836da8ae","Type":"ContainerStarted","Data":"eceb3c06d712707b71b887bb2059d84327ed026cb2a40bbc43c94040593009e3"} Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.349687 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.381895 4807 scope.go:117] "RemoveContainer" containerID="718476f3d7f59eab77a9aff9e3005dd34a8fc033c6e2af044c306b1afb284dd4" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.419059 4807 scope.go:117] "RemoveContainer" containerID="db6d3c9f2bdab5fcc8f703f696b1ee4f15e91c695eec6cfd6b2a82f26ebdfeed" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.429410 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.441648 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.456685 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/kube-state-metrics-0" podStartSLOduration=2.121108286 podStartE2EDuration="6.456665663s" podCreationTimestamp="2025-12-05 12:29:17 +0000 UTC" firstStartedPulling="2025-12-05 12:29:18.090874397 +0000 UTC m=+1387.584737666" lastFinishedPulling="2025-12-05 12:29:22.426431774 +0000 UTC m=+1391.920295043" observedRunningTime="2025-12-05 12:29:23.431807237 +0000 UTC m=+1392.925670526" watchObservedRunningTime="2025-12-05 12:29:23.456665663 +0000 UTC m=+1392.950528932" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.471619 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:29:23 crc kubenswrapper[4807]: E1205 12:29:23.472047 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-central-agent" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472062 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-central-agent" Dec 05 12:29:23 crc kubenswrapper[4807]: E1205 12:29:23.472083 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="sg-core" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472091 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="sg-core" Dec 05 12:29:23 crc kubenswrapper[4807]: E1205 12:29:23.472104 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-notification-agent" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472113 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-notification-agent" Dec 05 12:29:23 crc kubenswrapper[4807]: E1205 12:29:23.472138 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="proxy-httpd" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472146 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="proxy-httpd" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472335 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-central-agent" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472357 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="sg-core" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472379 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="ceilometer-notification-agent" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.472393 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" containerName="proxy-httpd" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.474209 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.476367 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.477448 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.477699 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.479755 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.496745 4807 scope.go:117] "RemoveContainer" containerID="a8e4a3acd7ca4b1a2ef63b40f11fe8b0b61ab3716db3a19f60d674aac9289e71" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643354 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643414 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-config-data\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643561 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-log-httpd\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643590 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643618 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlvpf\" (UniqueName: \"kubernetes.io/projected/142f37b8-bf78-4526-aa53-27c2aca9cfde-kube-api-access-xlvpf\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643653 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643690 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-run-httpd\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.643723 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-scripts\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.745562 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-config-data\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.745942 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-log-httpd\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.746029 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.746105 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlvpf\" (UniqueName: \"kubernetes.io/projected/142f37b8-bf78-4526-aa53-27c2aca9cfde-kube-api-access-xlvpf\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.746186 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.746276 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-run-httpd\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.746398 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-scripts\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.746516 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.747117 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-run-httpd\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.747678 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-log-httpd\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.749625 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.750127 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-config-data\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.750517 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-scripts\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.756220 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.758901 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.762988 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlvpf\" (UniqueName: \"kubernetes.io/projected/142f37b8-bf78-4526-aa53-27c2aca9cfde-kube-api-access-xlvpf\") pod \"ceilometer-0\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:23 crc kubenswrapper[4807]: I1205 12:29:23.816269 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.050360 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.057595 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.079434 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-operator-scripts\") pod \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.079491 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qslcf\" (UniqueName: \"kubernetes.io/projected/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-kube-api-access-qslcf\") pod \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\" (UID: \"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec\") " Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.079585 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhlwd\" (UniqueName: \"kubernetes.io/projected/3381f31f-572a-4e4f-aae5-d38d9a9253cd-kube-api-access-hhlwd\") pod \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.079658 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3381f31f-572a-4e4f-aae5-d38d9a9253cd-operator-scripts\") pod \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\" (UID: \"3381f31f-572a-4e4f-aae5-d38d9a9253cd\") " Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.079955 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3d9bd6b-3038-4cc3-b204-96b047cfe3ec" (UID: "a3d9bd6b-3038-4cc3-b204-96b047cfe3ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.080128 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.080984 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3381f31f-572a-4e4f-aae5-d38d9a9253cd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3381f31f-572a-4e4f-aae5-d38d9a9253cd" (UID: "3381f31f-572a-4e4f-aae5-d38d9a9253cd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.091605 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-kube-api-access-qslcf" (OuterVolumeSpecName: "kube-api-access-qslcf") pod "a3d9bd6b-3038-4cc3-b204-96b047cfe3ec" (UID: "a3d9bd6b-3038-4cc3-b204-96b047cfe3ec"). InnerVolumeSpecName "kube-api-access-qslcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.095544 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3381f31f-572a-4e4f-aae5-d38d9a9253cd-kube-api-access-hhlwd" (OuterVolumeSpecName: "kube-api-access-hhlwd") pod "3381f31f-572a-4e4f-aae5-d38d9a9253cd" (UID: "3381f31f-572a-4e4f-aae5-d38d9a9253cd"). InnerVolumeSpecName "kube-api-access-hhlwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.124053 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:29:25 crc kubenswrapper[4807]: W1205 12:29:25.128317 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod142f37b8_bf78_4526_aa53_27c2aca9cfde.slice/crio-c19c4db5d6c1241f14a26b27e402894e7c0fb67cf298d43a6aef96de8b698049 WatchSource:0}: Error finding container c19c4db5d6c1241f14a26b27e402894e7c0fb67cf298d43a6aef96de8b698049: Status 404 returned error can't find the container with id c19c4db5d6c1241f14a26b27e402894e7c0fb67cf298d43a6aef96de8b698049 Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.181082 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3381f31f-572a-4e4f-aae5-d38d9a9253cd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.181117 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qslcf\" (UniqueName: \"kubernetes.io/projected/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec-kube-api-access-qslcf\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.181129 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhlwd\" (UniqueName: \"kubernetes.io/projected/3381f31f-572a-4e4f-aae5-d38d9a9253cd-kube-api-access-hhlwd\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.245861 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98c406d3-6570-4956-ad01-fcf83a35f565" path="/var/lib/kubelet/pods/98c406d3-6570-4956-ad01-fcf83a35f565/volumes" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.369891 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5q5fh" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.369891 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-5q5fh" event={"ID":"a3d9bd6b-3038-4cc3-b204-96b047cfe3ec","Type":"ContainerDied","Data":"9423fb5fcb81b424653f1ee76c50981624a28272f38fd26cf21fd29a08416182"} Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.370082 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9423fb5fcb81b424653f1ee76c50981624a28272f38fd26cf21fd29a08416182" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.378724 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" event={"ID":"3381f31f-572a-4e4f-aae5-d38d9a9253cd","Type":"ContainerDied","Data":"52672ff8de092cfdaf33d78ee879de1cc9b55be2d15e741ff3069ce45c0ec094"} Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.378764 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52672ff8de092cfdaf33d78ee879de1cc9b55be2d15e741ff3069ce45c0ec094" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.378823 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r" Dec 05 12:29:25 crc kubenswrapper[4807]: I1205 12:29:25.391709 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerStarted","Data":"c19c4db5d6c1241f14a26b27e402894e7c0fb67cf298d43a6aef96de8b698049"} Dec 05 12:29:26 crc kubenswrapper[4807]: I1205 12:29:26.401743 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerStarted","Data":"c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9"} Dec 05 12:29:26 crc kubenswrapper[4807]: I1205 12:29:26.402275 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerStarted","Data":"f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1"} Dec 05 12:29:27 crc kubenswrapper[4807]: I1205 12:29:27.419201 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerStarted","Data":"832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938"} Dec 05 12:29:27 crc kubenswrapper[4807]: I1205 12:29:27.649476 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/kube-state-metrics-0" Dec 05 12:29:29 crc kubenswrapper[4807]: I1205 12:29:29.445000 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerStarted","Data":"920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b"} Dec 05 12:29:29 crc kubenswrapper[4807]: I1205 12:29:29.445411 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:29:29 crc kubenswrapper[4807]: I1205 12:29:29.474064 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=3.407169113 podStartE2EDuration="6.47404307s" podCreationTimestamp="2025-12-05 12:29:23 +0000 UTC" firstStartedPulling="2025-12-05 12:29:25.133232764 +0000 UTC m=+1394.627096033" lastFinishedPulling="2025-12-05 12:29:28.200106721 +0000 UTC m=+1397.693969990" observedRunningTime="2025-12-05 12:29:29.468129746 +0000 UTC m=+1398.961993015" watchObservedRunningTime="2025-12-05 12:29:29.47404307 +0000 UTC m=+1398.967906339" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.310083 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh"] Dec 05 12:29:30 crc kubenswrapper[4807]: E1205 12:29:30.310441 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3381f31f-572a-4e4f-aae5-d38d9a9253cd" containerName="mariadb-account-create-update" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.310463 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3381f31f-572a-4e4f-aae5-d38d9a9253cd" containerName="mariadb-account-create-update" Dec 05 12:29:30 crc kubenswrapper[4807]: E1205 12:29:30.310499 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d9bd6b-3038-4cc3-b204-96b047cfe3ec" containerName="mariadb-database-create" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.310509 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d9bd6b-3038-4cc3-b204-96b047cfe3ec" containerName="mariadb-database-create" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.310739 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d9bd6b-3038-4cc3-b204-96b047cfe3ec" containerName="mariadb-database-create" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.310767 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3381f31f-572a-4e4f-aae5-d38d9a9253cd" containerName="mariadb-account-create-update" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.311409 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.315412 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.315773 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-wp4km" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.337306 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh"] Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.463352 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.463478 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-db-sync-config-data\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.463604 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-config-data\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.463662 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljnkp\" (UniqueName: \"kubernetes.io/projected/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-kube-api-access-ljnkp\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.564949 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-db-sync-config-data\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.565233 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-config-data\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.565324 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljnkp\" (UniqueName: \"kubernetes.io/projected/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-kube-api-access-ljnkp\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.565421 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.572222 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-db-sync-config-data\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.572738 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-config-data\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.575108 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.589376 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljnkp\" (UniqueName: \"kubernetes.io/projected/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-kube-api-access-ljnkp\") pod \"watcher-kuttl-db-sync-4r5sh\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:30 crc kubenswrapper[4807]: I1205 12:29:30.629870 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:29:31 crc kubenswrapper[4807]: I1205 12:29:31.165373 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh"] Dec 05 12:29:31 crc kubenswrapper[4807]: I1205 12:29:31.464554 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" event={"ID":"c8b22391-1075-4580-8cb0-6b6ab59cfe5f","Type":"ContainerStarted","Data":"c3a10d1ce2e703c597fd0fa06d3f37d0de55bbc14dda0983ece433599dd4b11b"} Dec 05 12:29:45 crc kubenswrapper[4807]: E1205 12:29:45.623365 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.20:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Dec 05 12:29:45 crc kubenswrapper[4807]: E1205 12:29:45.624017 4807 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.20:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Dec 05 12:29:45 crc kubenswrapper[4807]: E1205 12:29:45.624159 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:watcher-kuttl-db-sync,Image:38.102.83.20:5001/podified-master-centos10/openstack-watcher-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/watcher/watcher.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:watcher-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljnkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-kuttl-db-sync-4r5sh_watcher-kuttl-default(c8b22391-1075-4580-8cb0-6b6ab59cfe5f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 12:29:45 crc kubenswrapper[4807]: E1205 12:29:45.625570 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-kuttl-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" podUID="c8b22391-1075-4580-8cb0-6b6ab59cfe5f" Dec 05 12:29:46 crc kubenswrapper[4807]: E1205 12:29:46.595815 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-kuttl-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.20:5001/podified-master-centos10/openstack-watcher-api:watcher_latest\\\"\"" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" podUID="c8b22391-1075-4580-8cb0-6b6ab59cfe5f" Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.465919 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.466449 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.466489 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.467132 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba4c03761578e8e0b9faf772d0c622df752a9d6198091d0d8315216f608b690f"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.467179 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://ba4c03761578e8e0b9faf772d0c622df752a9d6198091d0d8315216f608b690f" gracePeriod=600 Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.648064 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="ba4c03761578e8e0b9faf772d0c622df752a9d6198091d0d8315216f608b690f" exitCode=0 Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.648615 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"ba4c03761578e8e0b9faf772d0c622df752a9d6198091d0d8315216f608b690f"} Dec 05 12:29:52 crc kubenswrapper[4807]: I1205 12:29:52.648662 4807 scope.go:117] "RemoveContainer" containerID="b47bace47d16c688e6065026bd16a58454816df8e9f3811211e0c6bea14f1f68" Dec 05 12:29:52 crc kubenswrapper[4807]: E1205 12:29:52.687887 4807 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda779882f_2b7e_4ae0_addd_686fd4343bb6.slice/crio-ba4c03761578e8e0b9faf772d0c622df752a9d6198091d0d8315216f608b690f.scope\": RecentStats: unable to find data in memory cache]" Dec 05 12:29:53 crc kubenswrapper[4807]: I1205 12:29:53.659551 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e"} Dec 05 12:29:53 crc kubenswrapper[4807]: I1205 12:29:53.824395 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.009869 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-6ppt7" podUID="f7a1e998-d292-49df-88b2-832d0ccb1c83" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.009893 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-6ppt7" podUID="f7a1e998-d292-49df-88b2-832d0ccb1c83" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.385742 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl"] Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.387038 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.390026 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.399552 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl"] Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.400632 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.497852 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5164bf9e-2f8a-49d2-9b59-d79ca888f532-secret-volume\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.498041 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5164bf9e-2f8a-49d2-9b59-d79ca888f532-config-volume\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.498085 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mskmr\" (UniqueName: \"kubernetes.io/projected/5164bf9e-2f8a-49d2-9b59-d79ca888f532-kube-api-access-mskmr\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.599869 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5164bf9e-2f8a-49d2-9b59-d79ca888f532-config-volume\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.599988 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mskmr\" (UniqueName: \"kubernetes.io/projected/5164bf9e-2f8a-49d2-9b59-d79ca888f532-kube-api-access-mskmr\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.600066 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5164bf9e-2f8a-49d2-9b59-d79ca888f532-secret-volume\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.600861 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5164bf9e-2f8a-49d2-9b59-d79ca888f532-config-volume\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.612800 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5164bf9e-2f8a-49d2-9b59-d79ca888f532-secret-volume\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.636259 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mskmr\" (UniqueName: \"kubernetes.io/projected/5164bf9e-2f8a-49d2-9b59-d79ca888f532-kube-api-access-mskmr\") pod \"collect-profiles-29415630-ftdvl\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:00 crc kubenswrapper[4807]: I1205 12:30:00.712613 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:01 crc kubenswrapper[4807]: I1205 12:30:01.203975 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl"] Dec 05 12:30:01 crc kubenswrapper[4807]: I1205 12:30:01.734719 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" event={"ID":"5164bf9e-2f8a-49d2-9b59-d79ca888f532","Type":"ContainerStarted","Data":"29c50a3610f12714ac987228d8f8218afa1729c2f385f4e4adb34be54786a009"} Dec 05 12:30:01 crc kubenswrapper[4807]: I1205 12:30:01.734773 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" event={"ID":"5164bf9e-2f8a-49d2-9b59-d79ca888f532","Type":"ContainerStarted","Data":"07a1190511ed85bbbc4133a33be061f2b8268cc1605717283acb4f0282240d90"} Dec 05 12:30:02 crc kubenswrapper[4807]: I1205 12:30:02.743286 4807 generic.go:334] "Generic (PLEG): container finished" podID="5164bf9e-2f8a-49d2-9b59-d79ca888f532" containerID="29c50a3610f12714ac987228d8f8218afa1729c2f385f4e4adb34be54786a009" exitCode=0 Dec 05 12:30:02 crc kubenswrapper[4807]: I1205 12:30:02.743406 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" event={"ID":"5164bf9e-2f8a-49d2-9b59-d79ca888f532","Type":"ContainerDied","Data":"29c50a3610f12714ac987228d8f8218afa1729c2f385f4e4adb34be54786a009"} Dec 05 12:30:02 crc kubenswrapper[4807]: I1205 12:30:02.745341 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" event={"ID":"c8b22391-1075-4580-8cb0-6b6ab59cfe5f","Type":"ContainerStarted","Data":"703fb818c19be18bd95560b9e803aceee8efb580fec44fe5714bfff738e5773d"} Dec 05 12:30:02 crc kubenswrapper[4807]: I1205 12:30:02.789351 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" podStartSLOduration=2.036146714 podStartE2EDuration="32.789331764s" podCreationTimestamp="2025-12-05 12:29:30 +0000 UTC" firstStartedPulling="2025-12-05 12:29:31.174283848 +0000 UTC m=+1400.668147117" lastFinishedPulling="2025-12-05 12:30:01.927468898 +0000 UTC m=+1431.421332167" observedRunningTime="2025-12-05 12:30:02.785945551 +0000 UTC m=+1432.279808830" watchObservedRunningTime="2025-12-05 12:30:02.789331764 +0000 UTC m=+1432.283195033" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.081385 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.163193 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5164bf9e-2f8a-49d2-9b59-d79ca888f532-secret-volume\") pod \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.163263 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5164bf9e-2f8a-49d2-9b59-d79ca888f532-config-volume\") pod \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.163357 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mskmr\" (UniqueName: \"kubernetes.io/projected/5164bf9e-2f8a-49d2-9b59-d79ca888f532-kube-api-access-mskmr\") pod \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\" (UID: \"5164bf9e-2f8a-49d2-9b59-d79ca888f532\") " Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.164778 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5164bf9e-2f8a-49d2-9b59-d79ca888f532-config-volume" (OuterVolumeSpecName: "config-volume") pod "5164bf9e-2f8a-49d2-9b59-d79ca888f532" (UID: "5164bf9e-2f8a-49d2-9b59-d79ca888f532"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.170309 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5164bf9e-2f8a-49d2-9b59-d79ca888f532-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5164bf9e-2f8a-49d2-9b59-d79ca888f532" (UID: "5164bf9e-2f8a-49d2-9b59-d79ca888f532"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.170325 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5164bf9e-2f8a-49d2-9b59-d79ca888f532-kube-api-access-mskmr" (OuterVolumeSpecName: "kube-api-access-mskmr") pod "5164bf9e-2f8a-49d2-9b59-d79ca888f532" (UID: "5164bf9e-2f8a-49d2-9b59-d79ca888f532"). InnerVolumeSpecName "kube-api-access-mskmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.268476 4807 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5164bf9e-2f8a-49d2-9b59-d79ca888f532-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.268568 4807 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5164bf9e-2f8a-49d2-9b59-d79ca888f532-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.268584 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mskmr\" (UniqueName: \"kubernetes.io/projected/5164bf9e-2f8a-49d2-9b59-d79ca888f532-kube-api-access-mskmr\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.761330 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" event={"ID":"5164bf9e-2f8a-49d2-9b59-d79ca888f532","Type":"ContainerDied","Data":"07a1190511ed85bbbc4133a33be061f2b8268cc1605717283acb4f0282240d90"} Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.761365 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-ftdvl" Dec 05 12:30:04 crc kubenswrapper[4807]: I1205 12:30:04.761373 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07a1190511ed85bbbc4133a33be061f2b8268cc1605717283acb4f0282240d90" Dec 05 12:30:06 crc kubenswrapper[4807]: I1205 12:30:06.779540 4807 generic.go:334] "Generic (PLEG): container finished" podID="c8b22391-1075-4580-8cb0-6b6ab59cfe5f" containerID="703fb818c19be18bd95560b9e803aceee8efb580fec44fe5714bfff738e5773d" exitCode=0 Dec 05 12:30:06 crc kubenswrapper[4807]: I1205 12:30:06.779638 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" event={"ID":"c8b22391-1075-4580-8cb0-6b6ab59cfe5f","Type":"ContainerDied","Data":"703fb818c19be18bd95560b9e803aceee8efb580fec44fe5714bfff738e5773d"} Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.166316 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.226500 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljnkp\" (UniqueName: \"kubernetes.io/projected/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-kube-api-access-ljnkp\") pod \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.226670 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-db-sync-config-data\") pod \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.226697 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-config-data\") pod \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.226730 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-combined-ca-bundle\") pod \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\" (UID: \"c8b22391-1075-4580-8cb0-6b6ab59cfe5f\") " Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.232093 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c8b22391-1075-4580-8cb0-6b6ab59cfe5f" (UID: "c8b22391-1075-4580-8cb0-6b6ab59cfe5f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.233491 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-kube-api-access-ljnkp" (OuterVolumeSpecName: "kube-api-access-ljnkp") pod "c8b22391-1075-4580-8cb0-6b6ab59cfe5f" (UID: "c8b22391-1075-4580-8cb0-6b6ab59cfe5f"). InnerVolumeSpecName "kube-api-access-ljnkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.254971 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8b22391-1075-4580-8cb0-6b6ab59cfe5f" (UID: "c8b22391-1075-4580-8cb0-6b6ab59cfe5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.282787 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-config-data" (OuterVolumeSpecName: "config-data") pod "c8b22391-1075-4580-8cb0-6b6ab59cfe5f" (UID: "c8b22391-1075-4580-8cb0-6b6ab59cfe5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.330307 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljnkp\" (UniqueName: \"kubernetes.io/projected/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-kube-api-access-ljnkp\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.330627 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.330642 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.330655 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8b22391-1075-4580-8cb0-6b6ab59cfe5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.798478 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" event={"ID":"c8b22391-1075-4580-8cb0-6b6ab59cfe5f","Type":"ContainerDied","Data":"c3a10d1ce2e703c597fd0fa06d3f37d0de55bbc14dda0983ece433599dd4b11b"} Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.798547 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3a10d1ce2e703c597fd0fa06d3f37d0de55bbc14dda0983ece433599dd4b11b" Dec 05 12:30:08 crc kubenswrapper[4807]: I1205 12:30:08.798843 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.073266 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:09 crc kubenswrapper[4807]: E1205 12:30:09.073746 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5164bf9e-2f8a-49d2-9b59-d79ca888f532" containerName="collect-profiles" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.073768 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5164bf9e-2f8a-49d2-9b59-d79ca888f532" containerName="collect-profiles" Dec 05 12:30:09 crc kubenswrapper[4807]: E1205 12:30:09.073807 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b22391-1075-4580-8cb0-6b6ab59cfe5f" containerName="watcher-kuttl-db-sync" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.073818 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b22391-1075-4580-8cb0-6b6ab59cfe5f" containerName="watcher-kuttl-db-sync" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.074031 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b22391-1075-4580-8cb0-6b6ab59cfe5f" containerName="watcher-kuttl-db-sync" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.074063 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5164bf9e-2f8a-49d2-9b59-d79ca888f532" containerName="collect-profiles" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.074852 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.088728 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.089031 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-wp4km" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.144260 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.144311 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0cf97d-2172-4060-8199-6ff03e1a4401-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.144371 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.144443 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh2cb\" (UniqueName: \"kubernetes.io/projected/ce0cf97d-2172-4060-8199-6ff03e1a4401-kube-api-access-rh2cb\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.144472 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.160845 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.169578 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.174982 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.178500 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.196405 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.210435 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.212430 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.215716 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.229967 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.245878 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh2cb\" (UniqueName: \"kubernetes.io/projected/ce0cf97d-2172-4060-8199-6ff03e1a4401-kube-api-access-rh2cb\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.245945 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.245974 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.245999 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246081 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec5b882c-1f6f-41e7-b267-d9233c18c018-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246137 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36de6e97-6e8d-4ff2-bae8-45a264002f39-logs\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246157 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246180 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246202 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0cf97d-2172-4060-8199-6ff03e1a4401-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246247 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67f9b\" (UniqueName: \"kubernetes.io/projected/ec5b882c-1f6f-41e7-b267-d9233c18c018-kube-api-access-67f9b\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246279 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jnpc\" (UniqueName: \"kubernetes.io/projected/36de6e97-6e8d-4ff2-bae8-45a264002f39-kube-api-access-5jnpc\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246314 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246336 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.246367 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.249142 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0cf97d-2172-4060-8199-6ff03e1a4401-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.254513 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.254960 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.255307 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.276566 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh2cb\" (UniqueName: \"kubernetes.io/projected/ce0cf97d-2172-4060-8199-6ff03e1a4401-kube-api-access-rh2cb\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347547 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347612 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347656 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec5b882c-1f6f-41e7-b267-d9233c18c018-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347681 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36de6e97-6e8d-4ff2-bae8-45a264002f39-logs\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347701 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347726 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67f9b\" (UniqueName: \"kubernetes.io/projected/ec5b882c-1f6f-41e7-b267-d9233c18c018-kube-api-access-67f9b\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347742 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jnpc\" (UniqueName: \"kubernetes.io/projected/36de6e97-6e8d-4ff2-bae8-45a264002f39-kube-api-access-5jnpc\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347785 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.347815 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.348268 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36de6e97-6e8d-4ff2-bae8-45a264002f39-logs\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.348848 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec5b882c-1f6f-41e7-b267-d9233c18c018-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.351921 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.353160 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.353257 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.353990 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.354201 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.365142 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jnpc\" (UniqueName: \"kubernetes.io/projected/36de6e97-6e8d-4ff2-bae8-45a264002f39-kube-api-access-5jnpc\") pod \"watcher-kuttl-api-0\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.366590 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67f9b\" (UniqueName: \"kubernetes.io/projected/ec5b882c-1f6f-41e7-b267-d9233c18c018-kube-api-access-67f9b\") pod \"watcher-kuttl-applier-0\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.460466 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.493475 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.539856 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:09 crc kubenswrapper[4807]: I1205 12:30:09.945873 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:09 crc kubenswrapper[4807]: W1205 12:30:09.950763 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0cf97d_2172_4060_8199_6ff03e1a4401.slice/crio-b715677f8cd8c41561a2828f7c6461a7fdf886aa0705c1e6b277110cb08660d9 WatchSource:0}: Error finding container b715677f8cd8c41561a2828f7c6461a7fdf886aa0705c1e6b277110cb08660d9: Status 404 returned error can't find the container with id b715677f8cd8c41561a2828f7c6461a7fdf886aa0705c1e6b277110cb08660d9 Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.040584 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.111341 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:10 crc kubenswrapper[4807]: W1205 12:30:10.114234 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec5b882c_1f6f_41e7_b267_d9233c18c018.slice/crio-77519621bae22f40e20cf19561f74e96fa476c0ed28dcd5d47bff4e17fa2f953 WatchSource:0}: Error finding container 77519621bae22f40e20cf19561f74e96fa476c0ed28dcd5d47bff4e17fa2f953: Status 404 returned error can't find the container with id 77519621bae22f40e20cf19561f74e96fa476c0ed28dcd5d47bff4e17fa2f953 Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.818871 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"36de6e97-6e8d-4ff2-bae8-45a264002f39","Type":"ContainerStarted","Data":"0f8a686a5d0c75dd0634f89ca363f77558309b1add3d617015bdb9227a8e6b75"} Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.819467 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"36de6e97-6e8d-4ff2-bae8-45a264002f39","Type":"ContainerStarted","Data":"27f5a60abfaa0dc45005768ba8aaca84e93fa903456e06e70936227755c71f43"} Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.819484 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"36de6e97-6e8d-4ff2-bae8-45a264002f39","Type":"ContainerStarted","Data":"c61380c09d6cbe53734cb1f2f03ab2d74f48fe1694fc9c52cca74f545ee1ea1f"} Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.819905 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.822758 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ec5b882c-1f6f-41e7-b267-d9233c18c018","Type":"ContainerStarted","Data":"77519621bae22f40e20cf19561f74e96fa476c0ed28dcd5d47bff4e17fa2f953"} Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.824424 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ce0cf97d-2172-4060-8199-6ff03e1a4401","Type":"ContainerStarted","Data":"b715677f8cd8c41561a2828f7c6461a7fdf886aa0705c1e6b277110cb08660d9"} Dec 05 12:30:10 crc kubenswrapper[4807]: I1205 12:30:10.844051 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=1.844025834 podStartE2EDuration="1.844025834s" podCreationTimestamp="2025-12-05 12:30:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:30:10.842812124 +0000 UTC m=+1440.336675403" watchObservedRunningTime="2025-12-05 12:30:10.844025834 +0000 UTC m=+1440.337889103" Dec 05 12:30:12 crc kubenswrapper[4807]: I1205 12:30:12.847239 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ec5b882c-1f6f-41e7-b267-d9233c18c018","Type":"ContainerStarted","Data":"d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062"} Dec 05 12:30:12 crc kubenswrapper[4807]: I1205 12:30:12.850170 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ce0cf97d-2172-4060-8199-6ff03e1a4401","Type":"ContainerStarted","Data":"c614e145dae3ed22bd21e018269c306916ffc14f3e57cd61a91137e6b5b971e3"} Dec 05 12:30:12 crc kubenswrapper[4807]: I1205 12:30:12.850220 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:30:12 crc kubenswrapper[4807]: I1205 12:30:12.868588 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.026468139 podStartE2EDuration="3.868562196s" podCreationTimestamp="2025-12-05 12:30:09 +0000 UTC" firstStartedPulling="2025-12-05 12:30:10.124323622 +0000 UTC m=+1439.618186891" lastFinishedPulling="2025-12-05 12:30:11.966417679 +0000 UTC m=+1441.460280948" observedRunningTime="2025-12-05 12:30:12.864411565 +0000 UTC m=+1442.358274834" watchObservedRunningTime="2025-12-05 12:30:12.868562196 +0000 UTC m=+1442.362425495" Dec 05 12:30:12 crc kubenswrapper[4807]: I1205 12:30:12.888592 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.067448298 podStartE2EDuration="3.888555993s" podCreationTimestamp="2025-12-05 12:30:09 +0000 UTC" firstStartedPulling="2025-12-05 12:30:09.954447392 +0000 UTC m=+1439.448310661" lastFinishedPulling="2025-12-05 12:30:11.775555087 +0000 UTC m=+1441.269418356" observedRunningTime="2025-12-05 12:30:12.88392351 +0000 UTC m=+1442.377786889" watchObservedRunningTime="2025-12-05 12:30:12.888555993 +0000 UTC m=+1442.382419262" Dec 05 12:30:13 crc kubenswrapper[4807]: I1205 12:30:13.476240 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:14 crc kubenswrapper[4807]: I1205 12:30:14.493879 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:14 crc kubenswrapper[4807]: I1205 12:30:14.540637 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.461604 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.492838 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.494319 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.505802 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.540773 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.564884 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.909136 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.918108 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.939226 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:19 crc kubenswrapper[4807]: I1205 12:30:19.939815 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.600657 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.601359 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-central-agent" containerID="cri-o://f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1" gracePeriod=30 Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.601522 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="proxy-httpd" containerID="cri-o://920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b" gracePeriod=30 Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.601735 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-notification-agent" containerID="cri-o://c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9" gracePeriod=30 Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.601801 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="sg-core" containerID="cri-o://832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938" gracePeriod=30 Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.935901 4807 generic.go:334] "Generic (PLEG): container finished" podID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerID="920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b" exitCode=0 Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.935948 4807 generic.go:334] "Generic (PLEG): container finished" podID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerID="832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938" exitCode=2 Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.935971 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerDied","Data":"920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b"} Dec 05 12:30:22 crc kubenswrapper[4807]: I1205 12:30:22.935997 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerDied","Data":"832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938"} Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.450421 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh"] Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.475921 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-4r5sh"] Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.496992 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher2ca3-account-delete-7x4xw"] Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.498190 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.510315 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher2ca3-account-delete-7x4xw"] Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.518143 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.518361 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="ec5b882c-1f6f-41e7-b267-d9233c18c018" containerName="watcher-applier" containerID="cri-o://d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062" gracePeriod=30 Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.580093 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.580351 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="ce0cf97d-2172-4060-8199-6ff03e1a4401" containerName="watcher-decision-engine" containerID="cri-o://c614e145dae3ed22bd21e018269c306916ffc14f3e57cd61a91137e6b5b971e3" gracePeriod=30 Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.608313 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klzq2\" (UniqueName: \"kubernetes.io/projected/7e236387-b51d-4614-82fa-d7dcd9d18546-kube-api-access-klzq2\") pod \"watcher2ca3-account-delete-7x4xw\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.608385 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e236387-b51d-4614-82fa-d7dcd9d18546-operator-scripts\") pod \"watcher2ca3-account-delete-7x4xw\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.640146 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.640418 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-kuttl-api-log" containerID="cri-o://27f5a60abfaa0dc45005768ba8aaca84e93fa903456e06e70936227755c71f43" gracePeriod=30 Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.640644 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-api" containerID="cri-o://0f8a686a5d0c75dd0634f89ca363f77558309b1add3d617015bdb9227a8e6b75" gracePeriod=30 Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.710207 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klzq2\" (UniqueName: \"kubernetes.io/projected/7e236387-b51d-4614-82fa-d7dcd9d18546-kube-api-access-klzq2\") pod \"watcher2ca3-account-delete-7x4xw\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.710257 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e236387-b51d-4614-82fa-d7dcd9d18546-operator-scripts\") pod \"watcher2ca3-account-delete-7x4xw\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.711156 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e236387-b51d-4614-82fa-d7dcd9d18546-operator-scripts\") pod \"watcher2ca3-account-delete-7x4xw\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.742890 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klzq2\" (UniqueName: \"kubernetes.io/projected/7e236387-b51d-4614-82fa-d7dcd9d18546-kube-api-access-klzq2\") pod \"watcher2ca3-account-delete-7x4xw\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.817296 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.818802 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.127:3000/\": dial tcp 10.217.0.127:3000: connect: connection refused" Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.982353 4807 generic.go:334] "Generic (PLEG): container finished" podID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerID="f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1" exitCode=0 Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.982408 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerDied","Data":"f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1"} Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.994876 4807 generic.go:334] "Generic (PLEG): container finished" podID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerID="27f5a60abfaa0dc45005768ba8aaca84e93fa903456e06e70936227755c71f43" exitCode=143 Dec 05 12:30:23 crc kubenswrapper[4807]: I1205 12:30:23.994923 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"36de6e97-6e8d-4ff2-bae8-45a264002f39","Type":"ContainerDied","Data":"27f5a60abfaa0dc45005768ba8aaca84e93fa903456e06e70936227755c71f43"} Dec 05 12:30:24 crc kubenswrapper[4807]: I1205 12:30:24.327767 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher2ca3-account-delete-7x4xw"] Dec 05 12:30:24 crc kubenswrapper[4807]: E1205 12:30:24.543110 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:30:24 crc kubenswrapper[4807]: E1205 12:30:24.545128 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:30:24 crc kubenswrapper[4807]: E1205 12:30:24.547374 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:30:24 crc kubenswrapper[4807]: E1205 12:30:24.547489 4807 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="ec5b882c-1f6f-41e7-b267-d9233c18c018" containerName="watcher-applier" Dec 05 12:30:24 crc kubenswrapper[4807]: I1205 12:30:24.776662 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.0.131:9322/\": read tcp 10.217.0.2:44298->10.217.0.131:9322: read: connection reset by peer" Dec 05 12:30:24 crc kubenswrapper[4807]: I1205 12:30:24.776682 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.131:9322/\": read tcp 10.217.0.2:44294->10.217.0.131:9322: read: connection reset by peer" Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.005270 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" event={"ID":"7e236387-b51d-4614-82fa-d7dcd9d18546","Type":"ContainerStarted","Data":"714f34656212ba81b5b11ebf4ec2fdd4e513048f8881def632a5dd9892cd0144"} Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.005319 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" event={"ID":"7e236387-b51d-4614-82fa-d7dcd9d18546","Type":"ContainerStarted","Data":"e8b92edd5c8c655cc8905c0d216490032d924373cc3d474579363c9db9d2f9b0"} Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.023356 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" podStartSLOduration=2.023334394 podStartE2EDuration="2.023334394s" podCreationTimestamp="2025-12-05 12:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:30:25.018611078 +0000 UTC m=+1454.512474347" watchObservedRunningTime="2025-12-05 12:30:25.023334394 +0000 UTC m=+1454.517197663" Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.271961 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8b22391-1075-4580-8cb0-6b6ab59cfe5f" path="/var/lib/kubelet/pods/c8b22391-1075-4580-8cb0-6b6ab59cfe5f/volumes" Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.830682 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.976679 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-run-httpd\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.976750 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlvpf\" (UniqueName: \"kubernetes.io/projected/142f37b8-bf78-4526-aa53-27c2aca9cfde-kube-api-access-xlvpf\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.976858 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-config-data\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.976944 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-scripts\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.976984 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-sg-core-conf-yaml\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.977023 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-combined-ca-bundle\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.977049 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-ceilometer-tls-certs\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.977089 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-log-httpd\") pod \"142f37b8-bf78-4526-aa53-27c2aca9cfde\" (UID: \"142f37b8-bf78-4526-aa53-27c2aca9cfde\") " Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.978109 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.978164 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:30:25 crc kubenswrapper[4807]: I1205 12:30:25.986037 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/142f37b8-bf78-4526-aa53-27c2aca9cfde-kube-api-access-xlvpf" (OuterVolumeSpecName: "kube-api-access-xlvpf") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "kube-api-access-xlvpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:25.999853 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-scripts" (OuterVolumeSpecName: "scripts") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.015673 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.037282 4807 generic.go:334] "Generic (PLEG): container finished" podID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerID="0f8a686a5d0c75dd0634f89ca363f77558309b1add3d617015bdb9227a8e6b75" exitCode=0 Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.037343 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"36de6e97-6e8d-4ff2-bae8-45a264002f39","Type":"ContainerDied","Data":"0f8a686a5d0c75dd0634f89ca363f77558309b1add3d617015bdb9227a8e6b75"} Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.037373 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"36de6e97-6e8d-4ff2-bae8-45a264002f39","Type":"ContainerDied","Data":"c61380c09d6cbe53734cb1f2f03ab2d74f48fe1694fc9c52cca74f545ee1ea1f"} Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.037388 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c61380c09d6cbe53734cb1f2f03ab2d74f48fe1694fc9c52cca74f545ee1ea1f" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.040278 4807 generic.go:334] "Generic (PLEG): container finished" podID="7e236387-b51d-4614-82fa-d7dcd9d18546" containerID="714f34656212ba81b5b11ebf4ec2fdd4e513048f8881def632a5dd9892cd0144" exitCode=0 Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.040418 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" event={"ID":"7e236387-b51d-4614-82fa-d7dcd9d18546","Type":"ContainerDied","Data":"714f34656212ba81b5b11ebf4ec2fdd4e513048f8881def632a5dd9892cd0144"} Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.043206 4807 generic.go:334] "Generic (PLEG): container finished" podID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerID="c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9" exitCode=0 Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.043474 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerDied","Data":"c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9"} Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.043575 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"142f37b8-bf78-4526-aa53-27c2aca9cfde","Type":"ContainerDied","Data":"c19c4db5d6c1241f14a26b27e402894e7c0fb67cf298d43a6aef96de8b698049"} Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.043673 4807 scope.go:117] "RemoveContainer" containerID="920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.043850 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.076127 4807 generic.go:334] "Generic (PLEG): container finished" podID="ce0cf97d-2172-4060-8199-6ff03e1a4401" containerID="c614e145dae3ed22bd21e018269c306916ffc14f3e57cd61a91137e6b5b971e3" exitCode=0 Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.076381 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ce0cf97d-2172-4060-8199-6ff03e1a4401","Type":"ContainerDied","Data":"c614e145dae3ed22bd21e018269c306916ffc14f3e57cd61a91137e6b5b971e3"} Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.078447 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.078478 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.078492 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.078503 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/142f37b8-bf78-4526-aa53-27c2aca9cfde-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.078517 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlvpf\" (UniqueName: \"kubernetes.io/projected/142f37b8-bf78-4526-aa53-27c2aca9cfde-kube-api-access-xlvpf\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.097757 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.105812 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-config-data" (OuterVolumeSpecName: "config-data") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.117873 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "142f37b8-bf78-4526-aa53-27c2aca9cfde" (UID: "142f37b8-bf78-4526-aa53-27c2aca9cfde"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.136624 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.138703 4807 scope.go:117] "RemoveContainer" containerID="832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.145448 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.182386 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.182421 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.182434 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f37b8-bf78-4526-aa53-27c2aca9cfde-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.186300 4807 scope.go:117] "RemoveContainer" containerID="c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.209631 4807 scope.go:117] "RemoveContainer" containerID="f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.231798 4807 scope.go:117] "RemoveContainer" containerID="920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.232200 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b\": container with ID starting with 920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b not found: ID does not exist" containerID="920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.232262 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b"} err="failed to get container status \"920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b\": rpc error: code = NotFound desc = could not find container \"920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b\": container with ID starting with 920ac1af22503cf77882d9739c2c7be6f08d092caa1f080338b7a69b7881bf1b not found: ID does not exist" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.232297 4807 scope.go:117] "RemoveContainer" containerID="832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.232714 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938\": container with ID starting with 832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938 not found: ID does not exist" containerID="832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.232747 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938"} err="failed to get container status \"832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938\": rpc error: code = NotFound desc = could not find container \"832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938\": container with ID starting with 832629dad3f6f371ae5484bc9fc26c3a7c425852b9baaa00654e69609066c938 not found: ID does not exist" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.232767 4807 scope.go:117] "RemoveContainer" containerID="c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.233015 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9\": container with ID starting with c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9 not found: ID does not exist" containerID="c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.233044 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9"} err="failed to get container status \"c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9\": rpc error: code = NotFound desc = could not find container \"c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9\": container with ID starting with c59b35a1946fdb3e31184eae61086cd6af338f7c9b1825993fc8c893913268f9 not found: ID does not exist" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.233062 4807 scope.go:117] "RemoveContainer" containerID="f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.233270 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1\": container with ID starting with f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1 not found: ID does not exist" containerID="f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.233311 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1"} err="failed to get container status \"f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1\": rpc error: code = NotFound desc = could not find container \"f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1\": container with ID starting with f55d807528528d7e416646abf2d95b25793de73390f509a9f676dffb83d82fa1 not found: ID does not exist" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.283916 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-custom-prometheus-ca\") pod \"ce0cf97d-2172-4060-8199-6ff03e1a4401\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284087 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-config-data\") pod \"36de6e97-6e8d-4ff2-bae8-45a264002f39\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284151 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0cf97d-2172-4060-8199-6ff03e1a4401-logs\") pod \"ce0cf97d-2172-4060-8199-6ff03e1a4401\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284206 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-config-data\") pod \"ce0cf97d-2172-4060-8199-6ff03e1a4401\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284227 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36de6e97-6e8d-4ff2-bae8-45a264002f39-logs\") pod \"36de6e97-6e8d-4ff2-bae8-45a264002f39\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284265 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jnpc\" (UniqueName: \"kubernetes.io/projected/36de6e97-6e8d-4ff2-bae8-45a264002f39-kube-api-access-5jnpc\") pod \"36de6e97-6e8d-4ff2-bae8-45a264002f39\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284320 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh2cb\" (UniqueName: \"kubernetes.io/projected/ce0cf97d-2172-4060-8199-6ff03e1a4401-kube-api-access-rh2cb\") pod \"ce0cf97d-2172-4060-8199-6ff03e1a4401\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284404 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-combined-ca-bundle\") pod \"36de6e97-6e8d-4ff2-bae8-45a264002f39\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284557 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-custom-prometheus-ca\") pod \"36de6e97-6e8d-4ff2-bae8-45a264002f39\" (UID: \"36de6e97-6e8d-4ff2-bae8-45a264002f39\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.284603 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-combined-ca-bundle\") pod \"ce0cf97d-2172-4060-8199-6ff03e1a4401\" (UID: \"ce0cf97d-2172-4060-8199-6ff03e1a4401\") " Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.285840 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36de6e97-6e8d-4ff2-bae8-45a264002f39-logs" (OuterVolumeSpecName: "logs") pod "36de6e97-6e8d-4ff2-bae8-45a264002f39" (UID: "36de6e97-6e8d-4ff2-bae8-45a264002f39"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.285824 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce0cf97d-2172-4060-8199-6ff03e1a4401-logs" (OuterVolumeSpecName: "logs") pod "ce0cf97d-2172-4060-8199-6ff03e1a4401" (UID: "ce0cf97d-2172-4060-8199-6ff03e1a4401"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.287029 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0cf97d-2172-4060-8199-6ff03e1a4401-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.287071 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36de6e97-6e8d-4ff2-bae8-45a264002f39-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.291812 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce0cf97d-2172-4060-8199-6ff03e1a4401-kube-api-access-rh2cb" (OuterVolumeSpecName: "kube-api-access-rh2cb") pod "ce0cf97d-2172-4060-8199-6ff03e1a4401" (UID: "ce0cf97d-2172-4060-8199-6ff03e1a4401"). InnerVolumeSpecName "kube-api-access-rh2cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.291983 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36de6e97-6e8d-4ff2-bae8-45a264002f39-kube-api-access-5jnpc" (OuterVolumeSpecName: "kube-api-access-5jnpc") pod "36de6e97-6e8d-4ff2-bae8-45a264002f39" (UID: "36de6e97-6e8d-4ff2-bae8-45a264002f39"). InnerVolumeSpecName "kube-api-access-5jnpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.313516 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36de6e97-6e8d-4ff2-bae8-45a264002f39" (UID: "36de6e97-6e8d-4ff2-bae8-45a264002f39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.313715 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "ce0cf97d-2172-4060-8199-6ff03e1a4401" (UID: "ce0cf97d-2172-4060-8199-6ff03e1a4401"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.317748 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "36de6e97-6e8d-4ff2-bae8-45a264002f39" (UID: "36de6e97-6e8d-4ff2-bae8-45a264002f39"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.321050 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce0cf97d-2172-4060-8199-6ff03e1a4401" (UID: "ce0cf97d-2172-4060-8199-6ff03e1a4401"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.334081 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-config-data" (OuterVolumeSpecName: "config-data") pod "36de6e97-6e8d-4ff2-bae8-45a264002f39" (UID: "36de6e97-6e8d-4ff2-bae8-45a264002f39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.337801 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-config-data" (OuterVolumeSpecName: "config-data") pod "ce0cf97d-2172-4060-8199-6ff03e1a4401" (UID: "ce0cf97d-2172-4060-8199-6ff03e1a4401"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.376504 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.383067 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.388551 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.388818 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.388926 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.389255 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.389340 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36de6e97-6e8d-4ff2-bae8-45a264002f39-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.389422 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0cf97d-2172-4060-8199-6ff03e1a4401-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.389503 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jnpc\" (UniqueName: \"kubernetes.io/projected/36de6e97-6e8d-4ff2-bae8-45a264002f39-kube-api-access-5jnpc\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.389661 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh2cb\" (UniqueName: \"kubernetes.io/projected/ce0cf97d-2172-4060-8199-6ff03e1a4401-kube-api-access-rh2cb\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.406770 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.407139 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="sg-core" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407161 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="sg-core" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.407183 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-kuttl-api-log" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407191 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-kuttl-api-log" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.407234 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-central-agent" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407243 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-central-agent" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.407259 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="proxy-httpd" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407266 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="proxy-httpd" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.407286 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-notification-agent" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407293 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-notification-agent" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.407305 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0cf97d-2172-4060-8199-6ff03e1a4401" containerName="watcher-decision-engine" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407312 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0cf97d-2172-4060-8199-6ff03e1a4401" containerName="watcher-decision-engine" Dec 05 12:30:26 crc kubenswrapper[4807]: E1205 12:30:26.407326 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-api" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407332 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-api" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407475 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="sg-core" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407489 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-notification-agent" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407502 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="ceilometer-central-agent" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407553 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce0cf97d-2172-4060-8199-6ff03e1a4401" containerName="watcher-decision-engine" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407568 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-kuttl-api-log" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407592 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" containerName="proxy-httpd" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.407610 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" containerName="watcher-api" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.409965 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.412052 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.412290 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.412416 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.426431 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.491294 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.491952 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9jxq\" (UniqueName: \"kubernetes.io/projected/05e4fa60-da79-4105-a2cf-1276abaa9e98-kube-api-access-j9jxq\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.492045 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-log-httpd\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.492224 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-run-httpd\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.492441 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.492495 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.492576 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-scripts\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.492651 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-config-data\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.593932 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-log-httpd\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.593981 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-run-httpd\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594035 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594060 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594088 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-scripts\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594116 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-config-data\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594178 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594208 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9jxq\" (UniqueName: \"kubernetes.io/projected/05e4fa60-da79-4105-a2cf-1276abaa9e98-kube-api-access-j9jxq\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594629 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-run-httpd\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.594790 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-log-httpd\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.598516 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.599002 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.600191 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-scripts\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.600966 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-config-data\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.601587 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.614405 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9jxq\" (UniqueName: \"kubernetes.io/projected/05e4fa60-da79-4105-a2cf-1276abaa9e98-kube-api-access-j9jxq\") pod \"ceilometer-0\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.724666 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:26 crc kubenswrapper[4807]: I1205 12:30:26.952994 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.088719 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.089534 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ce0cf97d-2172-4060-8199-6ff03e1a4401","Type":"ContainerDied","Data":"b715677f8cd8c41561a2828f7c6461a7fdf886aa0705c1e6b277110cb08660d9"} Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.089587 4807 scope.go:117] "RemoveContainer" containerID="c614e145dae3ed22bd21e018269c306916ffc14f3e57cd61a91137e6b5b971e3" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.089738 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.131092 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.143256 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.150504 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.158341 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.187797 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:30:27 crc kubenswrapper[4807]: W1205 12:30:27.192513 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05e4fa60_da79_4105_a2cf_1276abaa9e98.slice/crio-c677cca58330f7ec0ff9b0ae2a318e36890c2ff64c873ef39619402c98f5c3fe WatchSource:0}: Error finding container c677cca58330f7ec0ff9b0ae2a318e36890c2ff64c873ef39619402c98f5c3fe: Status 404 returned error can't find the container with id c677cca58330f7ec0ff9b0ae2a318e36890c2ff64c873ef39619402c98f5c3fe Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.248125 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="142f37b8-bf78-4526-aa53-27c2aca9cfde" path="/var/lib/kubelet/pods/142f37b8-bf78-4526-aa53-27c2aca9cfde/volumes" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.252350 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36de6e97-6e8d-4ff2-bae8-45a264002f39" path="/var/lib/kubelet/pods/36de6e97-6e8d-4ff2-bae8-45a264002f39/volumes" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.253125 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce0cf97d-2172-4060-8199-6ff03e1a4401" path="/var/lib/kubelet/pods/ce0cf97d-2172-4060-8199-6ff03e1a4401/volumes" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.475804 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.615722 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e236387-b51d-4614-82fa-d7dcd9d18546-operator-scripts\") pod \"7e236387-b51d-4614-82fa-d7dcd9d18546\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.615856 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klzq2\" (UniqueName: \"kubernetes.io/projected/7e236387-b51d-4614-82fa-d7dcd9d18546-kube-api-access-klzq2\") pod \"7e236387-b51d-4614-82fa-d7dcd9d18546\" (UID: \"7e236387-b51d-4614-82fa-d7dcd9d18546\") " Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.617024 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e236387-b51d-4614-82fa-d7dcd9d18546-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e236387-b51d-4614-82fa-d7dcd9d18546" (UID: "7e236387-b51d-4614-82fa-d7dcd9d18546"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.621616 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e236387-b51d-4614-82fa-d7dcd9d18546-kube-api-access-klzq2" (OuterVolumeSpecName: "kube-api-access-klzq2") pod "7e236387-b51d-4614-82fa-d7dcd9d18546" (UID: "7e236387-b51d-4614-82fa-d7dcd9d18546"). InnerVolumeSpecName "kube-api-access-klzq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.717397 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e236387-b51d-4614-82fa-d7dcd9d18546-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:27 crc kubenswrapper[4807]: I1205 12:30:27.717444 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klzq2\" (UniqueName: \"kubernetes.io/projected/7e236387-b51d-4614-82fa-d7dcd9d18546-kube-api-access-klzq2\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.098158 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerStarted","Data":"c677cca58330f7ec0ff9b0ae2a318e36890c2ff64c873ef39619402c98f5c3fe"} Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.100018 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" event={"ID":"7e236387-b51d-4614-82fa-d7dcd9d18546","Type":"ContainerDied","Data":"e8b92edd5c8c655cc8905c0d216490032d924373cc3d474579363c9db9d2f9b0"} Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.100088 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8b92edd5c8c655cc8905c0d216490032d924373cc3d474579363c9db9d2f9b0" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.100036 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher2ca3-account-delete-7x4xw" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.102294 4807 generic.go:334] "Generic (PLEG): container finished" podID="ec5b882c-1f6f-41e7-b267-d9233c18c018" containerID="d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062" exitCode=0 Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.102356 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ec5b882c-1f6f-41e7-b267-d9233c18c018","Type":"ContainerDied","Data":"d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062"} Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.532109 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5q5fh"] Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.540418 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5q5fh"] Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.568537 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r"] Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.580387 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher2ca3-account-delete-7x4xw"] Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.583388 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.588577 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-2ca3-account-create-update-zzd7r"] Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.592334 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher2ca3-account-delete-7x4xw"] Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.732690 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-config-data\") pod \"ec5b882c-1f6f-41e7-b267-d9233c18c018\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.732859 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67f9b\" (UniqueName: \"kubernetes.io/projected/ec5b882c-1f6f-41e7-b267-d9233c18c018-kube-api-access-67f9b\") pod \"ec5b882c-1f6f-41e7-b267-d9233c18c018\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.732925 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec5b882c-1f6f-41e7-b267-d9233c18c018-logs\") pod \"ec5b882c-1f6f-41e7-b267-d9233c18c018\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.732957 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-combined-ca-bundle\") pod \"ec5b882c-1f6f-41e7-b267-d9233c18c018\" (UID: \"ec5b882c-1f6f-41e7-b267-d9233c18c018\") " Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.734555 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec5b882c-1f6f-41e7-b267-d9233c18c018-logs" (OuterVolumeSpecName: "logs") pod "ec5b882c-1f6f-41e7-b267-d9233c18c018" (UID: "ec5b882c-1f6f-41e7-b267-d9233c18c018"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.748762 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec5b882c-1f6f-41e7-b267-d9233c18c018-kube-api-access-67f9b" (OuterVolumeSpecName: "kube-api-access-67f9b") pod "ec5b882c-1f6f-41e7-b267-d9233c18c018" (UID: "ec5b882c-1f6f-41e7-b267-d9233c18c018"). InnerVolumeSpecName "kube-api-access-67f9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.757016 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec5b882c-1f6f-41e7-b267-d9233c18c018" (UID: "ec5b882c-1f6f-41e7-b267-d9233c18c018"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.783158 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-config-data" (OuterVolumeSpecName: "config-data") pod "ec5b882c-1f6f-41e7-b267-d9233c18c018" (UID: "ec5b882c-1f6f-41e7-b267-d9233c18c018"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.834907 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec5b882c-1f6f-41e7-b267-d9233c18c018-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.834943 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.834954 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec5b882c-1f6f-41e7-b267-d9233c18c018-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:28 crc kubenswrapper[4807]: I1205 12:30:28.834963 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67f9b\" (UniqueName: \"kubernetes.io/projected/ec5b882c-1f6f-41e7-b267-d9233c18c018-kube-api-access-67f9b\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.130833 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerStarted","Data":"891b042e34e110dbafab55b31116053a5b0da80b0c9c167fc0ee570ba70e50d9"} Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.133391 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ec5b882c-1f6f-41e7-b267-d9233c18c018","Type":"ContainerDied","Data":"77519621bae22f40e20cf19561f74e96fa476c0ed28dcd5d47bff4e17fa2f953"} Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.133428 4807 scope.go:117] "RemoveContainer" containerID="d0fb51027e93995bd074c30d526d8859b45e4ebe2cc90d7b9a0e1abc9eda8062" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.133547 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.195553 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.202511 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.244687 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3381f31f-572a-4e4f-aae5-d38d9a9253cd" path="/var/lib/kubelet/pods/3381f31f-572a-4e4f-aae5-d38d9a9253cd/volumes" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.245411 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e236387-b51d-4614-82fa-d7dcd9d18546" path="/var/lib/kubelet/pods/7e236387-b51d-4614-82fa-d7dcd9d18546/volumes" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.245975 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3d9bd6b-3038-4cc3-b204-96b047cfe3ec" path="/var/lib/kubelet/pods/a3d9bd6b-3038-4cc3-b204-96b047cfe3ec/volumes" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.247032 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec5b882c-1f6f-41e7-b267-d9233c18c018" path="/var/lib/kubelet/pods/ec5b882c-1f6f-41e7-b267-d9233c18c018/volumes" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.628366 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-85rbw"] Dec 05 12:30:29 crc kubenswrapper[4807]: E1205 12:30:29.628798 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e236387-b51d-4614-82fa-d7dcd9d18546" containerName="mariadb-account-delete" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.628819 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e236387-b51d-4614-82fa-d7dcd9d18546" containerName="mariadb-account-delete" Dec 05 12:30:29 crc kubenswrapper[4807]: E1205 12:30:29.628834 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5b882c-1f6f-41e7-b267-d9233c18c018" containerName="watcher-applier" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.628842 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5b882c-1f6f-41e7-b267-d9233c18c018" containerName="watcher-applier" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.629054 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e236387-b51d-4614-82fa-d7dcd9d18546" containerName="mariadb-account-delete" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.629084 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec5b882c-1f6f-41e7-b267-d9233c18c018" containerName="watcher-applier" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.629753 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.653666 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-37e0-account-create-update-94vgp"] Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.654722 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.656729 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.666928 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-85rbw"] Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.708318 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-37e0-account-create-update-94vgp"] Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.749427 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xw65\" (UniqueName: \"kubernetes.io/projected/f0c125c6-a97e-4163-8e9c-c56aec710a35-kube-api-access-6xw65\") pod \"watcher-db-create-85rbw\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.749712 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c125c6-a97e-4163-8e9c-c56aec710a35-operator-scripts\") pod \"watcher-db-create-85rbw\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.851305 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xw65\" (UniqueName: \"kubernetes.io/projected/f0c125c6-a97e-4163-8e9c-c56aec710a35-kube-api-access-6xw65\") pod \"watcher-db-create-85rbw\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.851383 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-operator-scripts\") pod \"watcher-37e0-account-create-update-94vgp\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.851437 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c125c6-a97e-4163-8e9c-c56aec710a35-operator-scripts\") pod \"watcher-db-create-85rbw\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.851483 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q7nz\" (UniqueName: \"kubernetes.io/projected/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-kube-api-access-2q7nz\") pod \"watcher-37e0-account-create-update-94vgp\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.852380 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c125c6-a97e-4163-8e9c-c56aec710a35-operator-scripts\") pod \"watcher-db-create-85rbw\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.872235 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xw65\" (UniqueName: \"kubernetes.io/projected/f0c125c6-a97e-4163-8e9c-c56aec710a35-kube-api-access-6xw65\") pod \"watcher-db-create-85rbw\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.947297 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.952149 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-operator-scripts\") pod \"watcher-37e0-account-create-update-94vgp\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.952296 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q7nz\" (UniqueName: \"kubernetes.io/projected/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-kube-api-access-2q7nz\") pod \"watcher-37e0-account-create-update-94vgp\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.953038 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-operator-scripts\") pod \"watcher-37e0-account-create-update-94vgp\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:29 crc kubenswrapper[4807]: I1205 12:30:29.971930 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q7nz\" (UniqueName: \"kubernetes.io/projected/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-kube-api-access-2q7nz\") pod \"watcher-37e0-account-create-update-94vgp\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:30 crc kubenswrapper[4807]: I1205 12:30:30.144769 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerStarted","Data":"95faee3384f11ba7efd6ffe507cf6d168624549dd569d6b97eb36f0fccc62596"} Dec 05 12:30:30 crc kubenswrapper[4807]: I1205 12:30:30.272071 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:30 crc kubenswrapper[4807]: I1205 12:30:30.422974 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-85rbw"] Dec 05 12:30:30 crc kubenswrapper[4807]: I1205 12:30:30.776702 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-37e0-account-create-update-94vgp"] Dec 05 12:30:31 crc kubenswrapper[4807]: I1205 12:30:31.159852 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" event={"ID":"3f02e2c4-fd4a-4607-b267-83afdcbeaa31","Type":"ContainerStarted","Data":"fe3fc9f58842ef225343f3b9d4ca0bd1b11ec153ccdab83ef892a49cd791a1dd"} Dec 05 12:30:31 crc kubenswrapper[4807]: I1205 12:30:31.164822 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-85rbw" event={"ID":"f0c125c6-a97e-4163-8e9c-c56aec710a35","Type":"ContainerStarted","Data":"ff6d04af820e28aa9d1bd63443c47294c91e8c22acb8cee9a004e6331a526adf"} Dec 05 12:30:32 crc kubenswrapper[4807]: I1205 12:30:32.172308 4807 generic.go:334] "Generic (PLEG): container finished" podID="f0c125c6-a97e-4163-8e9c-c56aec710a35" containerID="ae548c83f5026cc23d83db7dc1f21e86f531b12bc924d0513ec4095641a8c2dc" exitCode=0 Dec 05 12:30:32 crc kubenswrapper[4807]: I1205 12:30:32.172349 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-85rbw" event={"ID":"f0c125c6-a97e-4163-8e9c-c56aec710a35","Type":"ContainerDied","Data":"ae548c83f5026cc23d83db7dc1f21e86f531b12bc924d0513ec4095641a8c2dc"} Dec 05 12:30:32 crc kubenswrapper[4807]: I1205 12:30:32.174497 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerStarted","Data":"09d62b917688f706f0445175c6a63fbd21dfe5b331bba17d470f83240f62aae7"} Dec 05 12:30:32 crc kubenswrapper[4807]: I1205 12:30:32.175786 4807 generic.go:334] "Generic (PLEG): container finished" podID="3f02e2c4-fd4a-4607-b267-83afdcbeaa31" containerID="04b61f30585914600acb0b1bf3904adfc07b668dbedabec88fcc9a1c7fd0144f" exitCode=0 Dec 05 12:30:32 crc kubenswrapper[4807]: I1205 12:30:32.175818 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" event={"ID":"3f02e2c4-fd4a-4607-b267-83afdcbeaa31","Type":"ContainerDied","Data":"04b61f30585914600acb0b1bf3904adfc07b668dbedabec88fcc9a1c7fd0144f"} Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.189713 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerStarted","Data":"0a6db74cae0d9593cf675f4b80904d6f59d654edce002b35635b186d816e0af5"} Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.191365 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.190044 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="sg-core" containerID="cri-o://09d62b917688f706f0445175c6a63fbd21dfe5b331bba17d470f83240f62aae7" gracePeriod=30 Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.189892 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-central-agent" containerID="cri-o://891b042e34e110dbafab55b31116053a5b0da80b0c9c167fc0ee570ba70e50d9" gracePeriod=30 Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.190094 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-notification-agent" containerID="cri-o://95faee3384f11ba7efd6ffe507cf6d168624549dd569d6b97eb36f0fccc62596" gracePeriod=30 Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.190089 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="proxy-httpd" containerID="cri-o://0a6db74cae0d9593cf675f4b80904d6f59d654edce002b35635b186d816e0af5" gracePeriod=30 Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.235065 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.604963933 podStartE2EDuration="7.23504336s" podCreationTimestamp="2025-12-05 12:30:26 +0000 UTC" firstStartedPulling="2025-12-05 12:30:27.194944231 +0000 UTC m=+1456.688807500" lastFinishedPulling="2025-12-05 12:30:32.825023658 +0000 UTC m=+1462.318886927" observedRunningTime="2025-12-05 12:30:33.217235957 +0000 UTC m=+1462.711099226" watchObservedRunningTime="2025-12-05 12:30:33.23504336 +0000 UTC m=+1462.728906639" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.583202 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.613542 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.621402 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q7nz\" (UniqueName: \"kubernetes.io/projected/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-kube-api-access-2q7nz\") pod \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.621516 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-operator-scripts\") pod \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\" (UID: \"3f02e2c4-fd4a-4607-b267-83afdcbeaa31\") " Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.621621 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c125c6-a97e-4163-8e9c-c56aec710a35-operator-scripts\") pod \"f0c125c6-a97e-4163-8e9c-c56aec710a35\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.621654 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xw65\" (UniqueName: \"kubernetes.io/projected/f0c125c6-a97e-4163-8e9c-c56aec710a35-kube-api-access-6xw65\") pod \"f0c125c6-a97e-4163-8e9c-c56aec710a35\" (UID: \"f0c125c6-a97e-4163-8e9c-c56aec710a35\") " Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.623896 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f02e2c4-fd4a-4607-b267-83afdcbeaa31" (UID: "3f02e2c4-fd4a-4607-b267-83afdcbeaa31"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.624509 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0c125c6-a97e-4163-8e9c-c56aec710a35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0c125c6-a97e-4163-8e9c-c56aec710a35" (UID: "f0c125c6-a97e-4163-8e9c-c56aec710a35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.631702 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0c125c6-a97e-4163-8e9c-c56aec710a35-kube-api-access-6xw65" (OuterVolumeSpecName: "kube-api-access-6xw65") pod "f0c125c6-a97e-4163-8e9c-c56aec710a35" (UID: "f0c125c6-a97e-4163-8e9c-c56aec710a35"). InnerVolumeSpecName "kube-api-access-6xw65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.635239 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-kube-api-access-2q7nz" (OuterVolumeSpecName: "kube-api-access-2q7nz") pod "3f02e2c4-fd4a-4607-b267-83afdcbeaa31" (UID: "3f02e2c4-fd4a-4607-b267-83afdcbeaa31"). InnerVolumeSpecName "kube-api-access-2q7nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.723171 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q7nz\" (UniqueName: \"kubernetes.io/projected/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-kube-api-access-2q7nz\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.723218 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f02e2c4-fd4a-4607-b267-83afdcbeaa31-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.723231 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c125c6-a97e-4163-8e9c-c56aec710a35-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:33 crc kubenswrapper[4807]: I1205 12:30:33.723242 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xw65\" (UniqueName: \"kubernetes.io/projected/f0c125c6-a97e-4163-8e9c-c56aec710a35-kube-api-access-6xw65\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.200279 4807 generic.go:334] "Generic (PLEG): container finished" podID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerID="09d62b917688f706f0445175c6a63fbd21dfe5b331bba17d470f83240f62aae7" exitCode=2 Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.200318 4807 generic.go:334] "Generic (PLEG): container finished" podID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerID="95faee3384f11ba7efd6ffe507cf6d168624549dd569d6b97eb36f0fccc62596" exitCode=0 Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.200327 4807 generic.go:334] "Generic (PLEG): container finished" podID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerID="891b042e34e110dbafab55b31116053a5b0da80b0c9c167fc0ee570ba70e50d9" exitCode=0 Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.200350 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerDied","Data":"09d62b917688f706f0445175c6a63fbd21dfe5b331bba17d470f83240f62aae7"} Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.200384 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerDied","Data":"95faee3384f11ba7efd6ffe507cf6d168624549dd569d6b97eb36f0fccc62596"} Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.200397 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerDied","Data":"891b042e34e110dbafab55b31116053a5b0da80b0c9c167fc0ee570ba70e50d9"} Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.201958 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.201959 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-37e0-account-create-update-94vgp" event={"ID":"3f02e2c4-fd4a-4607-b267-83afdcbeaa31","Type":"ContainerDied","Data":"fe3fc9f58842ef225343f3b9d4ca0bd1b11ec153ccdab83ef892a49cd791a1dd"} Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.202033 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe3fc9f58842ef225343f3b9d4ca0bd1b11ec153ccdab83ef892a49cd791a1dd" Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.203576 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-85rbw" event={"ID":"f0c125c6-a97e-4163-8e9c-c56aec710a35","Type":"ContainerDied","Data":"ff6d04af820e28aa9d1bd63443c47294c91e8c22acb8cee9a004e6331a526adf"} Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.203604 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff6d04af820e28aa9d1bd63443c47294c91e8c22acb8cee9a004e6331a526adf" Dec 05 12:30:34 crc kubenswrapper[4807]: I1205 12:30:34.203639 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-85rbw" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.079724 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc"] Dec 05 12:30:35 crc kubenswrapper[4807]: E1205 12:30:35.080283 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f02e2c4-fd4a-4607-b267-83afdcbeaa31" containerName="mariadb-account-create-update" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.080295 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f02e2c4-fd4a-4607-b267-83afdcbeaa31" containerName="mariadb-account-create-update" Dec 05 12:30:35 crc kubenswrapper[4807]: E1205 12:30:35.080307 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0c125c6-a97e-4163-8e9c-c56aec710a35" containerName="mariadb-database-create" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.080312 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0c125c6-a97e-4163-8e9c-c56aec710a35" containerName="mariadb-database-create" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.080479 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0c125c6-a97e-4163-8e9c-c56aec710a35" containerName="mariadb-database-create" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.080502 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f02e2c4-fd4a-4607-b267-83afdcbeaa31" containerName="mariadb-account-create-update" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.081043 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.083150 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.083560 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-89prv" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.093035 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc"] Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.145076 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.145126 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4j2\" (UniqueName: \"kubernetes.io/projected/c8466ff3-477d-41b7-bcc4-0d6b39076535-kube-api-access-qr4j2\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.145209 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-db-sync-config-data\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.145254 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-config-data\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.246328 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-config-data\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.247041 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.247161 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4j2\" (UniqueName: \"kubernetes.io/projected/c8466ff3-477d-41b7-bcc4-0d6b39076535-kube-api-access-qr4j2\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.247371 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-db-sync-config-data\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.251036 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-db-sync-config-data\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.251193 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-config-data\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.251194 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.270730 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4j2\" (UniqueName: \"kubernetes.io/projected/c8466ff3-477d-41b7-bcc4-0d6b39076535-kube-api-access-qr4j2\") pod \"watcher-kuttl-db-sync-nb4qc\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.395940 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:35 crc kubenswrapper[4807]: I1205 12:30:35.914779 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc"] Dec 05 12:30:35 crc kubenswrapper[4807]: W1205 12:30:35.915743 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8466ff3_477d_41b7_bcc4_0d6b39076535.slice/crio-de1efbe8072c78c936fa629162a9d0cd495948b0c9eafe61bb353f37818c61ad WatchSource:0}: Error finding container de1efbe8072c78c936fa629162a9d0cd495948b0c9eafe61bb353f37818c61ad: Status 404 returned error can't find the container with id de1efbe8072c78c936fa629162a9d0cd495948b0c9eafe61bb353f37818c61ad Dec 05 12:30:36 crc kubenswrapper[4807]: I1205 12:30:36.220001 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" event={"ID":"c8466ff3-477d-41b7-bcc4-0d6b39076535","Type":"ContainerStarted","Data":"e58f694a858405dda6c07db5a11a37916e84824a7b44ad7a5c3f5b327ec6bdf8"} Dec 05 12:30:36 crc kubenswrapper[4807]: I1205 12:30:36.220314 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" event={"ID":"c8466ff3-477d-41b7-bcc4-0d6b39076535","Type":"ContainerStarted","Data":"de1efbe8072c78c936fa629162a9d0cd495948b0c9eafe61bb353f37818c61ad"} Dec 05 12:30:36 crc kubenswrapper[4807]: I1205 12:30:36.245265 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" podStartSLOduration=1.245248135 podStartE2EDuration="1.245248135s" podCreationTimestamp="2025-12-05 12:30:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:30:36.235652532 +0000 UTC m=+1465.729515811" watchObservedRunningTime="2025-12-05 12:30:36.245248135 +0000 UTC m=+1465.739111404" Dec 05 12:30:39 crc kubenswrapper[4807]: I1205 12:30:39.247227 4807 generic.go:334] "Generic (PLEG): container finished" podID="c8466ff3-477d-41b7-bcc4-0d6b39076535" containerID="e58f694a858405dda6c07db5a11a37916e84824a7b44ad7a5c3f5b327ec6bdf8" exitCode=0 Dec 05 12:30:39 crc kubenswrapper[4807]: I1205 12:30:39.247291 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" event={"ID":"c8466ff3-477d-41b7-bcc4-0d6b39076535","Type":"ContainerDied","Data":"e58f694a858405dda6c07db5a11a37916e84824a7b44ad7a5c3f5b327ec6bdf8"} Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.707553 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.751828 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-config-data\") pod \"c8466ff3-477d-41b7-bcc4-0d6b39076535\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.751981 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-combined-ca-bundle\") pod \"c8466ff3-477d-41b7-bcc4-0d6b39076535\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.752034 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr4j2\" (UniqueName: \"kubernetes.io/projected/c8466ff3-477d-41b7-bcc4-0d6b39076535-kube-api-access-qr4j2\") pod \"c8466ff3-477d-41b7-bcc4-0d6b39076535\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.752059 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-db-sync-config-data\") pod \"c8466ff3-477d-41b7-bcc4-0d6b39076535\" (UID: \"c8466ff3-477d-41b7-bcc4-0d6b39076535\") " Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.758719 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c8466ff3-477d-41b7-bcc4-0d6b39076535" (UID: "c8466ff3-477d-41b7-bcc4-0d6b39076535"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.774777 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8466ff3-477d-41b7-bcc4-0d6b39076535-kube-api-access-qr4j2" (OuterVolumeSpecName: "kube-api-access-qr4j2") pod "c8466ff3-477d-41b7-bcc4-0d6b39076535" (UID: "c8466ff3-477d-41b7-bcc4-0d6b39076535"). InnerVolumeSpecName "kube-api-access-qr4j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.782251 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8466ff3-477d-41b7-bcc4-0d6b39076535" (UID: "c8466ff3-477d-41b7-bcc4-0d6b39076535"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.812576 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-config-data" (OuterVolumeSpecName: "config-data") pod "c8466ff3-477d-41b7-bcc4-0d6b39076535" (UID: "c8466ff3-477d-41b7-bcc4-0d6b39076535"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.853676 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.853711 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.853723 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr4j2\" (UniqueName: \"kubernetes.io/projected/c8466ff3-477d-41b7-bcc4-0d6b39076535-kube-api-access-qr4j2\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:40 crc kubenswrapper[4807]: I1205 12:30:40.853733 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8466ff3-477d-41b7-bcc4-0d6b39076535-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.266659 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" event={"ID":"c8466ff3-477d-41b7-bcc4-0d6b39076535","Type":"ContainerDied","Data":"de1efbe8072c78c936fa629162a9d0cd495948b0c9eafe61bb353f37818c61ad"} Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.266720 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1efbe8072c78c936fa629162a9d0cd495948b0c9eafe61bb353f37818c61ad" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.266802 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.483515 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:41 crc kubenswrapper[4807]: E1205 12:30:41.483884 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8466ff3-477d-41b7-bcc4-0d6b39076535" containerName="watcher-kuttl-db-sync" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.483904 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8466ff3-477d-41b7-bcc4-0d6b39076535" containerName="watcher-kuttl-db-sync" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.484138 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8466ff3-477d-41b7-bcc4-0d6b39076535" containerName="watcher-kuttl-db-sync" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.485199 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.495507 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-89prv" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.498764 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.501429 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.567764 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.567786 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.568140 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.568258 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6v6x\" (UniqueName: \"kubernetes.io/projected/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-kube-api-access-b6v6x\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.568348 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.568466 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.568766 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.573941 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.582031 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.609061 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.610342 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.614211 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.642022 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.669855 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.669956 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.669980 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6v6x\" (UniqueName: \"kubernetes.io/projected/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-kube-api-access-b6v6x\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670005 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670038 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670055 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670068 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ade136a-7184-44d7-a0dd-8e079690bc0a-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670094 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79sx\" (UniqueName: \"kubernetes.io/projected/3ade136a-7184-44d7-a0dd-8e079690bc0a-kube-api-access-j79sx\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670127 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670161 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670198 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqrq2\" (UniqueName: \"kubernetes.io/projected/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-kube-api-access-gqrq2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670223 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670252 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.670274 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.671188 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.674423 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.676293 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.699949 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6v6x\" (UniqueName: \"kubernetes.io/projected/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-kube-api-access-b6v6x\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.699955 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772401 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772500 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ade136a-7184-44d7-a0dd-8e079690bc0a-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772582 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79sx\" (UniqueName: \"kubernetes.io/projected/3ade136a-7184-44d7-a0dd-8e079690bc0a-kube-api-access-j79sx\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772660 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772730 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772764 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqrq2\" (UniqueName: \"kubernetes.io/projected/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-kube-api-access-gqrq2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772828 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772934 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.772971 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.773156 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.774170 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ade136a-7184-44d7-a0dd-8e079690bc0a-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.777695 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.778047 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.779691 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.780436 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.781494 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.792032 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79sx\" (UniqueName: \"kubernetes.io/projected/3ade136a-7184-44d7-a0dd-8e079690bc0a-kube-api-access-j79sx\") pod \"watcher-kuttl-applier-0\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.792042 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqrq2\" (UniqueName: \"kubernetes.io/projected/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-kube-api-access-gqrq2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.807044 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.886686 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:41 crc kubenswrapper[4807]: I1205 12:30:41.934020 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:42 crc kubenswrapper[4807]: I1205 12:30:42.308475 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:42 crc kubenswrapper[4807]: I1205 12:30:42.444775 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:42 crc kubenswrapper[4807]: W1205 12:30:42.451709 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ade136a_7184_44d7_a0dd_8e079690bc0a.slice/crio-6d509d371195af3c2ad7741794f4a2caacaf1a2ead8f612fbc907057f0f1dcdf WatchSource:0}: Error finding container 6d509d371195af3c2ad7741794f4a2caacaf1a2ead8f612fbc907057f0f1dcdf: Status 404 returned error can't find the container with id 6d509d371195af3c2ad7741794f4a2caacaf1a2ead8f612fbc907057f0f1dcdf Dec 05 12:30:42 crc kubenswrapper[4807]: I1205 12:30:42.539162 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:42 crc kubenswrapper[4807]: W1205 12:30:42.550026 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d3e54a3_a1ee_4ae7_a4ed_775237ca354d.slice/crio-b045787c4a75d6b9c14b756c589104f052ca8112feb1bba4e012189a8fd1f6dc WatchSource:0}: Error finding container b045787c4a75d6b9c14b756c589104f052ca8112feb1bba4e012189a8fd1f6dc: Status 404 returned error can't find the container with id b045787c4a75d6b9c14b756c589104f052ca8112feb1bba4e012189a8fd1f6dc Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.305867 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d","Type":"ContainerStarted","Data":"b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5"} Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.306180 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d","Type":"ContainerStarted","Data":"b045787c4a75d6b9c14b756c589104f052ca8112feb1bba4e012189a8fd1f6dc"} Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.318206 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3ade136a-7184-44d7-a0dd-8e079690bc0a","Type":"ContainerStarted","Data":"8a126f4e260b944db863142be9b98a0847da1d056de89ded365169b185f85288"} Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.318256 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3ade136a-7184-44d7-a0dd-8e079690bc0a","Type":"ContainerStarted","Data":"6d509d371195af3c2ad7741794f4a2caacaf1a2ead8f612fbc907057f0f1dcdf"} Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.333410 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.333390158 podStartE2EDuration="2.333390158s" podCreationTimestamp="2025-12-05 12:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:30:43.32895122 +0000 UTC m=+1472.822814489" watchObservedRunningTime="2025-12-05 12:30:43.333390158 +0000 UTC m=+1472.827253427" Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.338801 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d","Type":"ContainerStarted","Data":"dd2398063e6497faca1c4c90e35c12a587727e56ebe81872e9d723e64ca7c738"} Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.338865 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d","Type":"ContainerStarted","Data":"895c464a110af8b6bfda5f57e19d62e855365ac139ac735eb2441c046f9af025"} Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.338878 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d","Type":"ContainerStarted","Data":"8e23b57c2934f23cb03b14e64d2e233ad28563a112316c9699d13a6aeefb7316"} Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.339823 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.364488 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.364467635 podStartE2EDuration="2.364467635s" podCreationTimestamp="2025-12-05 12:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:30:43.361845291 +0000 UTC m=+1472.855708560" watchObservedRunningTime="2025-12-05 12:30:43.364467635 +0000 UTC m=+1472.858330904" Dec 05 12:30:43 crc kubenswrapper[4807]: I1205 12:30:43.389995 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.389967987 podStartE2EDuration="2.389967987s" podCreationTimestamp="2025-12-05 12:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:30:43.383200432 +0000 UTC m=+1472.877063701" watchObservedRunningTime="2025-12-05 12:30:43.389967987 +0000 UTC m=+1472.883831246" Dec 05 12:30:45 crc kubenswrapper[4807]: I1205 12:30:45.353272 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:30:45 crc kubenswrapper[4807]: I1205 12:30:45.789820 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:46 crc kubenswrapper[4807]: I1205 12:30:46.808182 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:46 crc kubenswrapper[4807]: I1205 12:30:46.887404 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:51 crc kubenswrapper[4807]: I1205 12:30:51.807973 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:51 crc kubenswrapper[4807]: I1205 12:30:51.816461 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:51 crc kubenswrapper[4807]: I1205 12:30:51.887393 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:51 crc kubenswrapper[4807]: I1205 12:30:51.910946 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:51 crc kubenswrapper[4807]: I1205 12:30:51.934648 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:51 crc kubenswrapper[4807]: I1205 12:30:51.963454 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:52 crc kubenswrapper[4807]: I1205 12:30:52.404783 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:52 crc kubenswrapper[4807]: I1205 12:30:52.413019 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:52 crc kubenswrapper[4807]: I1205 12:30:52.437348 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:52 crc kubenswrapper[4807]: I1205 12:30:52.441001 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:30:56 crc kubenswrapper[4807]: I1205 12:30:56.734509 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.391370 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc"] Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.397596 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nb4qc"] Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.432309 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher37e0-account-delete-248vs"] Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.433849 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.452787 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher37e0-account-delete-248vs"] Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.492025 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.492349 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="3ade136a-7184-44d7-a0dd-8e079690bc0a" containerName="watcher-applier" containerID="cri-o://8a126f4e260b944db863142be9b98a0847da1d056de89ded365169b185f85288" gracePeriod=30 Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.542801 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.543093 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-kuttl-api-log" containerID="cri-o://895c464a110af8b6bfda5f57e19d62e855365ac139ac735eb2441c046f9af025" gracePeriod=30 Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.543274 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-api" containerID="cri-o://dd2398063e6497faca1c4c90e35c12a587727e56ebe81872e9d723e64ca7c738" gracePeriod=30 Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.546727 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6256de60-4a18-48cf-b005-43ceb54de04e-operator-scripts\") pod \"watcher37e0-account-delete-248vs\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.546906 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-958mf\" (UniqueName: \"kubernetes.io/projected/6256de60-4a18-48cf-b005-43ceb54de04e-kube-api-access-958mf\") pod \"watcher37e0-account-delete-248vs\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.562732 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.562951 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" containerName="watcher-decision-engine" containerID="cri-o://b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5" gracePeriod=30 Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.649176 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-958mf\" (UniqueName: \"kubernetes.io/projected/6256de60-4a18-48cf-b005-43ceb54de04e-kube-api-access-958mf\") pod \"watcher37e0-account-delete-248vs\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.649277 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6256de60-4a18-48cf-b005-43ceb54de04e-operator-scripts\") pod \"watcher37e0-account-delete-248vs\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.650509 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6256de60-4a18-48cf-b005-43ceb54de04e-operator-scripts\") pod \"watcher37e0-account-delete-248vs\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.671007 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-958mf\" (UniqueName: \"kubernetes.io/projected/6256de60-4a18-48cf-b005-43ceb54de04e-kube-api-access-958mf\") pod \"watcher37e0-account-delete-248vs\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:57 crc kubenswrapper[4807]: I1205 12:30:57.752792 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:30:58 crc kubenswrapper[4807]: I1205 12:30:58.443849 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher37e0-account-delete-248vs"] Dec 05 12:30:58 crc kubenswrapper[4807]: I1205 12:30:58.468978 4807 generic.go:334] "Generic (PLEG): container finished" podID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerID="895c464a110af8b6bfda5f57e19d62e855365ac139ac735eb2441c046f9af025" exitCode=143 Dec 05 12:30:58 crc kubenswrapper[4807]: I1205 12:30:58.469040 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d","Type":"ContainerDied","Data":"895c464a110af8b6bfda5f57e19d62e855365ac139ac735eb2441c046f9af025"} Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.249115 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8466ff3-477d-41b7-bcc4-0d6b39076535" path="/var/lib/kubelet/pods/c8466ff3-477d-41b7-bcc4-0d6b39076535/volumes" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.487346 4807 generic.go:334] "Generic (PLEG): container finished" podID="3ade136a-7184-44d7-a0dd-8e079690bc0a" containerID="8a126f4e260b944db863142be9b98a0847da1d056de89ded365169b185f85288" exitCode=0 Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.487447 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3ade136a-7184-44d7-a0dd-8e079690bc0a","Type":"ContainerDied","Data":"8a126f4e260b944db863142be9b98a0847da1d056de89ded365169b185f85288"} Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.492993 4807 generic.go:334] "Generic (PLEG): container finished" podID="6256de60-4a18-48cf-b005-43ceb54de04e" containerID="c34a04d8ae1613d29f01bc9c4d3b791f4d026f8d7a19a3994535c8cd62445371" exitCode=0 Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.493499 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" event={"ID":"6256de60-4a18-48cf-b005-43ceb54de04e","Type":"ContainerDied","Data":"c34a04d8ae1613d29f01bc9c4d3b791f4d026f8d7a19a3994535c8cd62445371"} Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.493581 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" event={"ID":"6256de60-4a18-48cf-b005-43ceb54de04e","Type":"ContainerStarted","Data":"eef81a7baf372eaff08de60f2ca353c1bf8518e22cab883936ad9dc4c77625ba"} Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.497556 4807 generic.go:334] "Generic (PLEG): container finished" podID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerID="dd2398063e6497faca1c4c90e35c12a587727e56ebe81872e9d723e64ca7c738" exitCode=0 Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.497591 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d","Type":"ContainerDied","Data":"dd2398063e6497faca1c4c90e35c12a587727e56ebe81872e9d723e64ca7c738"} Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.565474 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.690178 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ade136a-7184-44d7-a0dd-8e079690bc0a-logs\") pod \"3ade136a-7184-44d7-a0dd-8e079690bc0a\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.690240 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-config-data\") pod \"3ade136a-7184-44d7-a0dd-8e079690bc0a\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.690281 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j79sx\" (UniqueName: \"kubernetes.io/projected/3ade136a-7184-44d7-a0dd-8e079690bc0a-kube-api-access-j79sx\") pod \"3ade136a-7184-44d7-a0dd-8e079690bc0a\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.690345 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-combined-ca-bundle\") pod \"3ade136a-7184-44d7-a0dd-8e079690bc0a\" (UID: \"3ade136a-7184-44d7-a0dd-8e079690bc0a\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.690457 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ade136a-7184-44d7-a0dd-8e079690bc0a-logs" (OuterVolumeSpecName: "logs") pod "3ade136a-7184-44d7-a0dd-8e079690bc0a" (UID: "3ade136a-7184-44d7-a0dd-8e079690bc0a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.690856 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ade136a-7184-44d7-a0dd-8e079690bc0a-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.698049 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ade136a-7184-44d7-a0dd-8e079690bc0a-kube-api-access-j79sx" (OuterVolumeSpecName: "kube-api-access-j79sx") pod "3ade136a-7184-44d7-a0dd-8e079690bc0a" (UID: "3ade136a-7184-44d7-a0dd-8e079690bc0a"). InnerVolumeSpecName "kube-api-access-j79sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.700629 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.754985 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ade136a-7184-44d7-a0dd-8e079690bc0a" (UID: "3ade136a-7184-44d7-a0dd-8e079690bc0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.769422 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-config-data" (OuterVolumeSpecName: "config-data") pod "3ade136a-7184-44d7-a0dd-8e079690bc0a" (UID: "3ade136a-7184-44d7-a0dd-8e079690bc0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.792149 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.792223 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j79sx\" (UniqueName: \"kubernetes.io/projected/3ade136a-7184-44d7-a0dd-8e079690bc0a-kube-api-access-j79sx\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.792247 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade136a-7184-44d7-a0dd-8e079690bc0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.893405 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-custom-prometheus-ca\") pod \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.893506 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-logs\") pod \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.893591 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6v6x\" (UniqueName: \"kubernetes.io/projected/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-kube-api-access-b6v6x\") pod \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.893717 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-combined-ca-bundle\") pod \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.893825 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-config-data\") pod \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\" (UID: \"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d\") " Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.894099 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-logs" (OuterVolumeSpecName: "logs") pod "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" (UID: "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.894401 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.896931 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-kube-api-access-b6v6x" (OuterVolumeSpecName: "kube-api-access-b6v6x") pod "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" (UID: "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d"). InnerVolumeSpecName "kube-api-access-b6v6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.917641 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" (UID: "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.927605 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" (UID: "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.947070 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-config-data" (OuterVolumeSpecName: "config-data") pod "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" (UID: "cadd4fa2-7fd1-4542-93e1-17b5194c5e0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.995775 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.996152 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.996165 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:59 crc kubenswrapper[4807]: I1205 12:30:59.996179 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6v6x\" (UniqueName: \"kubernetes.io/projected/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d-kube-api-access-b6v6x\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.416733 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.509026 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.509043 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"cadd4fa2-7fd1-4542-93e1-17b5194c5e0d","Type":"ContainerDied","Data":"8e23b57c2934f23cb03b14e64d2e233ad28563a112316c9699d13a6aeefb7316"} Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.509101 4807 scope.go:117] "RemoveContainer" containerID="dd2398063e6497faca1c4c90e35c12a587727e56ebe81872e9d723e64ca7c738" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.511492 4807 generic.go:334] "Generic (PLEG): container finished" podID="3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" containerID="b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5" exitCode=0 Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.511582 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d","Type":"ContainerDied","Data":"b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5"} Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.511635 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d","Type":"ContainerDied","Data":"b045787c4a75d6b9c14b756c589104f052ca8112feb1bba4e012189a8fd1f6dc"} Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.511812 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.515177 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.516644 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3ade136a-7184-44d7-a0dd-8e079690bc0a","Type":"ContainerDied","Data":"6d509d371195af3c2ad7741794f4a2caacaf1a2ead8f612fbc907057f0f1dcdf"} Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.542374 4807 scope.go:117] "RemoveContainer" containerID="895c464a110af8b6bfda5f57e19d62e855365ac139ac735eb2441c046f9af025" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.553227 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.566118 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.583624 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.604427 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqrq2\" (UniqueName: \"kubernetes.io/projected/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-kube-api-access-gqrq2\") pod \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.604480 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-config-data\") pod \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.604588 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-logs\") pod \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.604653 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-custom-prometheus-ca\") pod \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.604669 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-combined-ca-bundle\") pod \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\" (UID: \"3d3e54a3-a1ee-4ae7-a4ed-775237ca354d\") " Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.606725 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.607419 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-logs" (OuterVolumeSpecName: "logs") pod "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" (UID: "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.621728 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-kube-api-access-gqrq2" (OuterVolumeSpecName: "kube-api-access-gqrq2") pod "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" (UID: "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d"). InnerVolumeSpecName "kube-api-access-gqrq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.623879 4807 scope.go:117] "RemoveContainer" containerID="b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.633174 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" (UID: "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.659018 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-config-data" (OuterVolumeSpecName: "config-data") pod "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" (UID: "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.662679 4807 scope.go:117] "RemoveContainer" containerID="b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5" Dec 05 12:31:00 crc kubenswrapper[4807]: E1205 12:31:00.663045 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5\": container with ID starting with b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5 not found: ID does not exist" containerID="b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.663084 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5"} err="failed to get container status \"b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5\": rpc error: code = NotFound desc = could not find container \"b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5\": container with ID starting with b0b70ce78420536f6d8f623ac996d419d6a7f72e3812b280d5bc5281302e24e5 not found: ID does not exist" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.663109 4807 scope.go:117] "RemoveContainer" containerID="8a126f4e260b944db863142be9b98a0847da1d056de89ded365169b185f85288" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.669184 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" (UID: "3d3e54a3-a1ee-4ae7-a4ed-775237ca354d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.708492 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.708564 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.708581 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.708596 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqrq2\" (UniqueName: \"kubernetes.io/projected/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-kube-api-access-gqrq2\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.708636 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.850064 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.857122 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:00 crc kubenswrapper[4807]: I1205 12:31:00.930940 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.012148 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-958mf\" (UniqueName: \"kubernetes.io/projected/6256de60-4a18-48cf-b005-43ceb54de04e-kube-api-access-958mf\") pod \"6256de60-4a18-48cf-b005-43ceb54de04e\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.012303 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6256de60-4a18-48cf-b005-43ceb54de04e-operator-scripts\") pod \"6256de60-4a18-48cf-b005-43ceb54de04e\" (UID: \"6256de60-4a18-48cf-b005-43ceb54de04e\") " Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.013070 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6256de60-4a18-48cf-b005-43ceb54de04e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6256de60-4a18-48cf-b005-43ceb54de04e" (UID: "6256de60-4a18-48cf-b005-43ceb54de04e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.030830 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6256de60-4a18-48cf-b005-43ceb54de04e-kube-api-access-958mf" (OuterVolumeSpecName: "kube-api-access-958mf") pod "6256de60-4a18-48cf-b005-43ceb54de04e" (UID: "6256de60-4a18-48cf-b005-43ceb54de04e"). InnerVolumeSpecName "kube-api-access-958mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.113597 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6256de60-4a18-48cf-b005-43ceb54de04e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.113636 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-958mf\" (UniqueName: \"kubernetes.io/projected/6256de60-4a18-48cf-b005-43ceb54de04e-kube-api-access-958mf\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.246391 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ade136a-7184-44d7-a0dd-8e079690bc0a" path="/var/lib/kubelet/pods/3ade136a-7184-44d7-a0dd-8e079690bc0a/volumes" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.247238 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" path="/var/lib/kubelet/pods/3d3e54a3-a1ee-4ae7-a4ed-775237ca354d/volumes" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.247959 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" path="/var/lib/kubelet/pods/cadd4fa2-7fd1-4542-93e1-17b5194c5e0d/volumes" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.527556 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" event={"ID":"6256de60-4a18-48cf-b005-43ceb54de04e","Type":"ContainerDied","Data":"eef81a7baf372eaff08de60f2ca353c1bf8518e22cab883936ad9dc4c77625ba"} Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.527604 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eef81a7baf372eaff08de60f2ca353c1bf8518e22cab883936ad9dc4c77625ba" Dec 05 12:31:01 crc kubenswrapper[4807]: I1205 12:31:01.527662 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher37e0-account-delete-248vs" Dec 05 12:31:02 crc kubenswrapper[4807]: I1205 12:31:02.460159 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-85rbw"] Dec 05 12:31:02 crc kubenswrapper[4807]: I1205 12:31:02.467866 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-85rbw"] Dec 05 12:31:02 crc kubenswrapper[4807]: I1205 12:31:02.480590 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher37e0-account-delete-248vs"] Dec 05 12:31:02 crc kubenswrapper[4807]: I1205 12:31:02.486749 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-37e0-account-create-update-94vgp"] Dec 05 12:31:02 crc kubenswrapper[4807]: I1205 12:31:02.493727 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher37e0-account-delete-248vs"] Dec 05 12:31:02 crc kubenswrapper[4807]: I1205 12:31:02.502740 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-37e0-account-create-update-94vgp"] Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.257034 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f02e2c4-fd4a-4607-b267-83afdcbeaa31" path="/var/lib/kubelet/pods/3f02e2c4-fd4a-4607-b267-83afdcbeaa31/volumes" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.257540 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6256de60-4a18-48cf-b005-43ceb54de04e" path="/var/lib/kubelet/pods/6256de60-4a18-48cf-b005-43ceb54de04e/volumes" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.258062 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0c125c6-a97e-4163-8e9c-c56aec710a35" path="/var/lib/kubelet/pods/f0c125c6-a97e-4163-8e9c-c56aec710a35/volumes" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.544509 4807 generic.go:334] "Generic (PLEG): container finished" podID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerID="0a6db74cae0d9593cf675f4b80904d6f59d654edce002b35635b186d816e0af5" exitCode=137 Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.544566 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerDied","Data":"0a6db74cae0d9593cf675f4b80904d6f59d654edce002b35635b186d816e0af5"} Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.544883 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"05e4fa60-da79-4105-a2cf-1276abaa9e98","Type":"ContainerDied","Data":"c677cca58330f7ec0ff9b0ae2a318e36890c2ff64c873ef39619402c98f5c3fe"} Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.544904 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c677cca58330f7ec0ff9b0ae2a318e36890c2ff64c873ef39619402c98f5c3fe" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.590095 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.653921 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-scripts\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654023 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-config-data\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654062 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9jxq\" (UniqueName: \"kubernetes.io/projected/05e4fa60-da79-4105-a2cf-1276abaa9e98-kube-api-access-j9jxq\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654134 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-sg-core-conf-yaml\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654168 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-combined-ca-bundle\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654193 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-run-httpd\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654223 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-ceilometer-tls-certs\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654250 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-log-httpd\") pod \"05e4fa60-da79-4105-a2cf-1276abaa9e98\" (UID: \"05e4fa60-da79-4105-a2cf-1276abaa9e98\") " Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654778 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.654934 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.663063 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05e4fa60-da79-4105-a2cf-1276abaa9e98-kube-api-access-j9jxq" (OuterVolumeSpecName: "kube-api-access-j9jxq") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "kube-api-access-j9jxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.666676 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-scripts" (OuterVolumeSpecName: "scripts") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.685351 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.727572 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.734793 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.755912 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9jxq\" (UniqueName: \"kubernetes.io/projected/05e4fa60-da79-4105-a2cf-1276abaa9e98-kube-api-access-j9jxq\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.755952 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.755962 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.755971 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.755981 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.755990 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/05e4fa60-da79-4105-a2cf-1276abaa9e98-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.755997 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.777470 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-config-data" (OuterVolumeSpecName: "config-data") pod "05e4fa60-da79-4105-a2cf-1276abaa9e98" (UID: "05e4fa60-da79-4105-a2cf-1276abaa9e98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:03 crc kubenswrapper[4807]: I1205 12:31:03.857016 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4fa60-da79-4105-a2cf-1276abaa9e98-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.456737 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-lpj9k"] Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457250 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ade136a-7184-44d7-a0dd-8e079690bc0a" containerName="watcher-applier" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457261 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ade136a-7184-44d7-a0dd-8e079690bc0a" containerName="watcher-applier" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457276 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6256de60-4a18-48cf-b005-43ceb54de04e" containerName="mariadb-account-delete" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457281 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="6256de60-4a18-48cf-b005-43ceb54de04e" containerName="mariadb-account-delete" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457292 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-kuttl-api-log" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457299 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-kuttl-api-log" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457308 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-notification-agent" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457314 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-notification-agent" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457323 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="sg-core" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457328 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="sg-core" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457339 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="proxy-httpd" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457345 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="proxy-httpd" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457354 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" containerName="watcher-decision-engine" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457360 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" containerName="watcher-decision-engine" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457372 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-api" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457378 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-api" Dec 05 12:31:04 crc kubenswrapper[4807]: E1205 12:31:04.457393 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-central-agent" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457399 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-central-agent" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457544 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d3e54a3-a1ee-4ae7-a4ed-775237ca354d" containerName="watcher-decision-engine" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457553 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-kuttl-api-log" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457563 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="cadd4fa2-7fd1-4542-93e1-17b5194c5e0d" containerName="watcher-api" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457574 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="6256de60-4a18-48cf-b005-43ceb54de04e" containerName="mariadb-account-delete" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457585 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-notification-agent" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457594 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="ceilometer-central-agent" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457604 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="proxy-httpd" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457609 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ade136a-7184-44d7-a0dd-8e079690bc0a" containerName="watcher-applier" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.457616 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" containerName="sg-core" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.458075 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.463954 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-operator-scripts\") pod \"watcher-db-create-lpj9k\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.464004 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-595bf\" (UniqueName: \"kubernetes.io/projected/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-kube-api-access-595bf\") pod \"watcher-db-create-lpj9k\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.486595 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lpj9k"] Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.551267 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.560409 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-b626-account-create-update-5v8wr"] Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.562871 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.564890 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.565417 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-operator-scripts\") pod \"watcher-b626-account-create-update-5v8wr\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.565461 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-operator-scripts\") pod \"watcher-db-create-lpj9k\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.565507 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-595bf\" (UniqueName: \"kubernetes.io/projected/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-kube-api-access-595bf\") pod \"watcher-db-create-lpj9k\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.565561 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqkpr\" (UniqueName: \"kubernetes.io/projected/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-kube-api-access-mqkpr\") pod \"watcher-b626-account-create-update-5v8wr\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.566167 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-operator-scripts\") pod \"watcher-db-create-lpj9k\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.568520 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-b626-account-create-update-5v8wr"] Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.597253 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-595bf\" (UniqueName: \"kubernetes.io/projected/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-kube-api-access-595bf\") pod \"watcher-db-create-lpj9k\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.615355 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.659691 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.672015 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqkpr\" (UniqueName: \"kubernetes.io/projected/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-kube-api-access-mqkpr\") pod \"watcher-b626-account-create-update-5v8wr\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.672152 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-operator-scripts\") pod \"watcher-b626-account-create-update-5v8wr\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.672285 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.695648 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-operator-scripts\") pod \"watcher-b626-account-create-update-5v8wr\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.699798 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.705564 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.705719 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.712104 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.713021 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.734292 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqkpr\" (UniqueName: \"kubernetes.io/projected/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-kube-api-access-mqkpr\") pod \"watcher-b626-account-create-update-5v8wr\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.773389 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-scripts\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.783960 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.784078 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.784213 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-log-httpd\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.784357 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.784503 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-config-data\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.784639 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-run-httpd\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.785312 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.785879 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qwwt\" (UniqueName: \"kubernetes.io/projected/876342fb-b348-49f2-bea7-42a4a055c3db-kube-api-access-4qwwt\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891324 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-config-data\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891716 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-run-httpd\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891741 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qwwt\" (UniqueName: \"kubernetes.io/projected/876342fb-b348-49f2-bea7-42a4a055c3db-kube-api-access-4qwwt\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891797 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-scripts\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891832 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891848 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891877 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-log-httpd\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.891898 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.894476 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.895721 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-run-httpd\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.898374 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-log-httpd\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.899729 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-config-data\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.902072 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-scripts\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.902575 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.903622 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.903633 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:04 crc kubenswrapper[4807]: I1205 12:31:04.929637 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qwwt\" (UniqueName: \"kubernetes.io/projected/876342fb-b348-49f2-bea7-42a4a055c3db-kube-api-access-4qwwt\") pod \"ceilometer-0\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:05 crc kubenswrapper[4807]: I1205 12:31:05.061321 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:05 crc kubenswrapper[4807]: I1205 12:31:05.248843 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05e4fa60-da79-4105-a2cf-1276abaa9e98" path="/var/lib/kubelet/pods/05e4fa60-da79-4105-a2cf-1276abaa9e98/volumes" Dec 05 12:31:05 crc kubenswrapper[4807]: I1205 12:31:05.405599 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lpj9k"] Dec 05 12:31:05 crc kubenswrapper[4807]: I1205 12:31:05.511232 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:05 crc kubenswrapper[4807]: W1205 12:31:05.518117 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod876342fb_b348_49f2_bea7_42a4a055c3db.slice/crio-114a6106604e71bd5f8ffa8ddf43c53dede54c755f202a83f80c6849ede2143d WatchSource:0}: Error finding container 114a6106604e71bd5f8ffa8ddf43c53dede54c755f202a83f80c6849ede2143d: Status 404 returned error can't find the container with id 114a6106604e71bd5f8ffa8ddf43c53dede54c755f202a83f80c6849ede2143d Dec 05 12:31:05 crc kubenswrapper[4807]: I1205 12:31:05.557136 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-b626-account-create-update-5v8wr"] Dec 05 12:31:05 crc kubenswrapper[4807]: I1205 12:31:05.572265 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerStarted","Data":"114a6106604e71bd5f8ffa8ddf43c53dede54c755f202a83f80c6849ede2143d"} Dec 05 12:31:05 crc kubenswrapper[4807]: I1205 12:31:05.574028 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-lpj9k" event={"ID":"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c","Type":"ContainerStarted","Data":"062a9e94d65d335622957bc20e683b1cfb56a3f385600e00e598064848b2d5fc"} Dec 05 12:31:06 crc kubenswrapper[4807]: I1205 12:31:06.592136 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerStarted","Data":"05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83"} Dec 05 12:31:06 crc kubenswrapper[4807]: I1205 12:31:06.594445 4807 generic.go:334] "Generic (PLEG): container finished" podID="fe61e51f-15fe-4b89-a025-4d3949ecbbc2" containerID="ce648a9c894f943215c7b189f27414c93f74984371f651e939c8047db0313c24" exitCode=0 Dec 05 12:31:06 crc kubenswrapper[4807]: I1205 12:31:06.594496 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" event={"ID":"fe61e51f-15fe-4b89-a025-4d3949ecbbc2","Type":"ContainerDied","Data":"ce648a9c894f943215c7b189f27414c93f74984371f651e939c8047db0313c24"} Dec 05 12:31:06 crc kubenswrapper[4807]: I1205 12:31:06.594516 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" event={"ID":"fe61e51f-15fe-4b89-a025-4d3949ecbbc2","Type":"ContainerStarted","Data":"7d1e1855e3cb09e8a1c174396b130610d05631ed2782bb816b5d639f6358862b"} Dec 05 12:31:06 crc kubenswrapper[4807]: I1205 12:31:06.598379 4807 generic.go:334] "Generic (PLEG): container finished" podID="3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c" containerID="a510ff0d47aa643abbacdc68049e138893a0fc057413878ac2825e532904adf4" exitCode=0 Dec 05 12:31:06 crc kubenswrapper[4807]: I1205 12:31:06.598458 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-lpj9k" event={"ID":"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c","Type":"ContainerDied","Data":"a510ff0d47aa643abbacdc68049e138893a0fc057413878ac2825e532904adf4"} Dec 05 12:31:07 crc kubenswrapper[4807]: I1205 12:31:07.606476 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerStarted","Data":"6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585"} Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.019576 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.026133 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.049924 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-595bf\" (UniqueName: \"kubernetes.io/projected/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-kube-api-access-595bf\") pod \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.049989 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqkpr\" (UniqueName: \"kubernetes.io/projected/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-kube-api-access-mqkpr\") pod \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.050060 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-operator-scripts\") pod \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\" (UID: \"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c\") " Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.050087 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-operator-scripts\") pod \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\" (UID: \"fe61e51f-15fe-4b89-a025-4d3949ecbbc2\") " Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.051148 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe61e51f-15fe-4b89-a025-4d3949ecbbc2" (UID: "fe61e51f-15fe-4b89-a025-4d3949ecbbc2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.051796 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c" (UID: "3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.055946 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-kube-api-access-mqkpr" (OuterVolumeSpecName: "kube-api-access-mqkpr") pod "fe61e51f-15fe-4b89-a025-4d3949ecbbc2" (UID: "fe61e51f-15fe-4b89-a025-4d3949ecbbc2"). InnerVolumeSpecName "kube-api-access-mqkpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.085245 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-kube-api-access-595bf" (OuterVolumeSpecName: "kube-api-access-595bf") pod "3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c" (UID: "3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c"). InnerVolumeSpecName "kube-api-access-595bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.151569 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqkpr\" (UniqueName: \"kubernetes.io/projected/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-kube-api-access-mqkpr\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.151602 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.151611 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe61e51f-15fe-4b89-a025-4d3949ecbbc2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.151620 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-595bf\" (UniqueName: \"kubernetes.io/projected/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c-kube-api-access-595bf\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.620799 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerStarted","Data":"f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d"} Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.623515 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" event={"ID":"fe61e51f-15fe-4b89-a025-4d3949ecbbc2","Type":"ContainerDied","Data":"7d1e1855e3cb09e8a1c174396b130610d05631ed2782bb816b5d639f6358862b"} Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.623664 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d1e1855e3cb09e8a1c174396b130610d05631ed2782bb816b5d639f6358862b" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.623820 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b626-account-create-update-5v8wr" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.625589 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-lpj9k" event={"ID":"3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c","Type":"ContainerDied","Data":"062a9e94d65d335622957bc20e683b1cfb56a3f385600e00e598064848b2d5fc"} Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.625617 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="062a9e94d65d335622957bc20e683b1cfb56a3f385600e00e598064848b2d5fc" Dec 05 12:31:08 crc kubenswrapper[4807]: I1205 12:31:08.625675 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lpj9k" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.644953 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerStarted","Data":"c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc"} Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.645361 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.677544 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.326311205 podStartE2EDuration="5.677458735s" podCreationTimestamp="2025-12-05 12:31:04 +0000 UTC" firstStartedPulling="2025-12-05 12:31:05.526155291 +0000 UTC m=+1495.020018560" lastFinishedPulling="2025-12-05 12:31:08.877302811 +0000 UTC m=+1498.371166090" observedRunningTime="2025-12-05 12:31:09.671254435 +0000 UTC m=+1499.165117704" watchObservedRunningTime="2025-12-05 12:31:09.677458735 +0000 UTC m=+1499.171322004" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.970998 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz"] Dec 05 12:31:09 crc kubenswrapper[4807]: E1205 12:31:09.972725 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c" containerName="mariadb-database-create" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.972846 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c" containerName="mariadb-database-create" Dec 05 12:31:09 crc kubenswrapper[4807]: E1205 12:31:09.972952 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe61e51f-15fe-4b89-a025-4d3949ecbbc2" containerName="mariadb-account-create-update" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.973046 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe61e51f-15fe-4b89-a025-4d3949ecbbc2" containerName="mariadb-account-create-update" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.973433 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c" containerName="mariadb-database-create" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.973550 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe61e51f-15fe-4b89-a025-4d3949ecbbc2" containerName="mariadb-account-create-update" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.974359 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.978180 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-rmwpm" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.978495 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.979202 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-config-data\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.979255 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl8fq\" (UniqueName: \"kubernetes.io/projected/66b5ee11-c239-4de9-8ac6-0750131b4927-kube-api-access-nl8fq\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.979284 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.979303 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-db-sync-config-data\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:09 crc kubenswrapper[4807]: I1205 12:31:09.984353 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz"] Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.080594 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-config-data\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.080674 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl8fq\" (UniqueName: \"kubernetes.io/projected/66b5ee11-c239-4de9-8ac6-0750131b4927-kube-api-access-nl8fq\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.080724 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.080753 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-db-sync-config-data\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.088128 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-db-sync-config-data\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.099865 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.102301 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-config-data\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.102580 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl8fq\" (UniqueName: \"kubernetes.io/projected/66b5ee11-c239-4de9-8ac6-0750131b4927-kube-api-access-nl8fq\") pod \"watcher-kuttl-db-sync-mlbzz\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.334215 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:10 crc kubenswrapper[4807]: I1205 12:31:10.833100 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz"] Dec 05 12:31:10 crc kubenswrapper[4807]: W1205 12:31:10.836590 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66b5ee11_c239_4de9_8ac6_0750131b4927.slice/crio-0d975b5c356321841613ef9d7af926df7c29b83446bd6f2cb735d89d3df88d48 WatchSource:0}: Error finding container 0d975b5c356321841613ef9d7af926df7c29b83446bd6f2cb735d89d3df88d48: Status 404 returned error can't find the container with id 0d975b5c356321841613ef9d7af926df7c29b83446bd6f2cb735d89d3df88d48 Dec 05 12:31:11 crc kubenswrapper[4807]: I1205 12:31:11.663009 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" event={"ID":"66b5ee11-c239-4de9-8ac6-0750131b4927","Type":"ContainerStarted","Data":"9ed390e0375a47c630701ff52555a9bc9fba7bb942482a71b08a94bec8bb3a38"} Dec 05 12:31:11 crc kubenswrapper[4807]: I1205 12:31:11.663048 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" event={"ID":"66b5ee11-c239-4de9-8ac6-0750131b4927","Type":"ContainerStarted","Data":"0d975b5c356321841613ef9d7af926df7c29b83446bd6f2cb735d89d3df88d48"} Dec 05 12:31:11 crc kubenswrapper[4807]: I1205 12:31:11.687642 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" podStartSLOduration=2.687615374 podStartE2EDuration="2.687615374s" podCreationTimestamp="2025-12-05 12:31:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:11.678697307 +0000 UTC m=+1501.172560566" watchObservedRunningTime="2025-12-05 12:31:11.687615374 +0000 UTC m=+1501.181478643" Dec 05 12:31:13 crc kubenswrapper[4807]: I1205 12:31:13.680582 4807 generic.go:334] "Generic (PLEG): container finished" podID="66b5ee11-c239-4de9-8ac6-0750131b4927" containerID="9ed390e0375a47c630701ff52555a9bc9fba7bb942482a71b08a94bec8bb3a38" exitCode=0 Dec 05 12:31:13 crc kubenswrapper[4807]: I1205 12:31:13.680685 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" event={"ID":"66b5ee11-c239-4de9-8ac6-0750131b4927","Type":"ContainerDied","Data":"9ed390e0375a47c630701ff52555a9bc9fba7bb942482a71b08a94bec8bb3a38"} Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.096498 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.263408 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-combined-ca-bundle\") pod \"66b5ee11-c239-4de9-8ac6-0750131b4927\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.263457 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-db-sync-config-data\") pod \"66b5ee11-c239-4de9-8ac6-0750131b4927\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.263498 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-config-data\") pod \"66b5ee11-c239-4de9-8ac6-0750131b4927\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.263615 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl8fq\" (UniqueName: \"kubernetes.io/projected/66b5ee11-c239-4de9-8ac6-0750131b4927-kube-api-access-nl8fq\") pod \"66b5ee11-c239-4de9-8ac6-0750131b4927\" (UID: \"66b5ee11-c239-4de9-8ac6-0750131b4927\") " Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.268896 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "66b5ee11-c239-4de9-8ac6-0750131b4927" (UID: "66b5ee11-c239-4de9-8ac6-0750131b4927"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.269737 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b5ee11-c239-4de9-8ac6-0750131b4927-kube-api-access-nl8fq" (OuterVolumeSpecName: "kube-api-access-nl8fq") pod "66b5ee11-c239-4de9-8ac6-0750131b4927" (UID: "66b5ee11-c239-4de9-8ac6-0750131b4927"). InnerVolumeSpecName "kube-api-access-nl8fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.291823 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66b5ee11-c239-4de9-8ac6-0750131b4927" (UID: "66b5ee11-c239-4de9-8ac6-0750131b4927"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.335326 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-config-data" (OuterVolumeSpecName: "config-data") pod "66b5ee11-c239-4de9-8ac6-0750131b4927" (UID: "66b5ee11-c239-4de9-8ac6-0750131b4927"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.365999 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.366053 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.366067 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b5ee11-c239-4de9-8ac6-0750131b4927-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.366080 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl8fq\" (UniqueName: \"kubernetes.io/projected/66b5ee11-c239-4de9-8ac6-0750131b4927-kube-api-access-nl8fq\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.695936 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" event={"ID":"66b5ee11-c239-4de9-8ac6-0750131b4927","Type":"ContainerDied","Data":"0d975b5c356321841613ef9d7af926df7c29b83446bd6f2cb735d89d3df88d48"} Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.695989 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d975b5c356321841613ef9d7af926df7c29b83446bd6f2cb735d89d3df88d48" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.696077 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.984417 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:15 crc kubenswrapper[4807]: E1205 12:31:15.984910 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b5ee11-c239-4de9-8ac6-0750131b4927" containerName="watcher-kuttl-db-sync" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.984936 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b5ee11-c239-4de9-8ac6-0750131b4927" containerName="watcher-kuttl-db-sync" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.985165 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b5ee11-c239-4de9-8ac6-0750131b4927" containerName="watcher-kuttl-db-sync" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.986357 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.989307 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.990774 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-rmwpm" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.991034 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.992034 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.993151 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:15 crc kubenswrapper[4807]: I1205 12:31:15.993153 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.003564 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.016351 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.040278 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.081535 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.081589 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.081609 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.081632 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdmpr\" (UniqueName: \"kubernetes.io/projected/4d510c3b-7662-4467-95d9-e28afb3d3e75-kube-api-access-wdmpr\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.081666 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.081689 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.081714 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d510c3b-7662-4467-95d9-e28afb3d3e75-logs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.124271 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.125309 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.129742 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.137287 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183498 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d510c3b-7662-4467-95d9-e28afb3d3e75-logs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183686 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5vx7\" (UniqueName: \"kubernetes.io/projected/d535f51e-aece-4220-a465-650d815d6a12-kube-api-access-r5vx7\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183738 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183756 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183787 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183807 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183820 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.183844 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdmpr\" (UniqueName: \"kubernetes.io/projected/4d510c3b-7662-4467-95d9-e28afb3d3e75-kube-api-access-wdmpr\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.184036 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d510c3b-7662-4467-95d9-e28afb3d3e75-logs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.184194 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.184304 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.184399 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d535f51e-aece-4220-a465-650d815d6a12-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.189298 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.189670 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.190205 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.193639 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.200563 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.201757 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdmpr\" (UniqueName: \"kubernetes.io/projected/4d510c3b-7662-4467-95d9-e28afb3d3e75-kube-api-access-wdmpr\") pod \"watcher-kuttl-api-0\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285426 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d535f51e-aece-4220-a465-650d815d6a12-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285490 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91b04dec-7037-469e-a32f-26aa6ef58d81-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285510 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5vx7\" (UniqueName: \"kubernetes.io/projected/d535f51e-aece-4220-a465-650d815d6a12-kube-api-access-r5vx7\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285550 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285587 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285602 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285641 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285661 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp2ht\" (UniqueName: \"kubernetes.io/projected/91b04dec-7037-469e-a32f-26aa6ef58d81-kube-api-access-tp2ht\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.285701 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.286641 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d535f51e-aece-4220-a465-650d815d6a12-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.288689 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.288981 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.306247 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5vx7\" (UniqueName: \"kubernetes.io/projected/d535f51e-aece-4220-a465-650d815d6a12-kube-api-access-r5vx7\") pod \"watcher-kuttl-applier-0\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.320202 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.359782 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.389408 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp2ht\" (UniqueName: \"kubernetes.io/projected/91b04dec-7037-469e-a32f-26aa6ef58d81-kube-api-access-tp2ht\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.389492 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.389568 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.389634 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91b04dec-7037-469e-a32f-26aa6ef58d81-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.389658 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.390246 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91b04dec-7037-469e-a32f-26aa6ef58d81-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.392991 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.396101 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.410946 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.411456 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp2ht\" (UniqueName: \"kubernetes.io/projected/91b04dec-7037-469e-a32f-26aa6ef58d81-kube-api-access-tp2ht\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.444247 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.823164 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:16 crc kubenswrapper[4807]: I1205 12:31:16.952890 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.125726 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.715642 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"91b04dec-7037-469e-a32f-26aa6ef58d81","Type":"ContainerStarted","Data":"0369ec80be57c6cb3ec1d477a47962495b62f837262bf38d299f2569f826322c"} Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.715933 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"91b04dec-7037-469e-a32f-26aa6ef58d81","Type":"ContainerStarted","Data":"be301c896545720b73dead96128195206ed663cd75fe4a997999a9393870982a"} Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.717790 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"d535f51e-aece-4220-a465-650d815d6a12","Type":"ContainerStarted","Data":"94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b"} Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.717820 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"d535f51e-aece-4220-a465-650d815d6a12","Type":"ContainerStarted","Data":"79633ae00afbcde37575d17407bae0dd3405e33f624232bb4a604edd46888946"} Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.719780 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"4d510c3b-7662-4467-95d9-e28afb3d3e75","Type":"ContainerStarted","Data":"d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd"} Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.719810 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"4d510c3b-7662-4467-95d9-e28afb3d3e75","Type":"ContainerStarted","Data":"bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045"} Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.719821 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"4d510c3b-7662-4467-95d9-e28afb3d3e75","Type":"ContainerStarted","Data":"928adceeb8f13ba23a59db0a1de0f8a67339747e1a8fb529ee96a7c1309cac19"} Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.720746 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.750563 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=1.75054261 podStartE2EDuration="1.75054261s" podCreationTimestamp="2025-12-05 12:31:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:17.746099102 +0000 UTC m=+1507.239962371" watchObservedRunningTime="2025-12-05 12:31:17.75054261 +0000 UTC m=+1507.244405879" Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.785008 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.784981438 podStartE2EDuration="2.784981438s" podCreationTimestamp="2025-12-05 12:31:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:17.779210258 +0000 UTC m=+1507.273073527" watchObservedRunningTime="2025-12-05 12:31:17.784981438 +0000 UTC m=+1507.278844707" Dec 05 12:31:17 crc kubenswrapper[4807]: I1205 12:31:17.856518 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.856491328 podStartE2EDuration="2.856491328s" podCreationTimestamp="2025-12-05 12:31:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:17.833329234 +0000 UTC m=+1507.327192503" watchObservedRunningTime="2025-12-05 12:31:17.856491328 +0000 UTC m=+1507.350354587" Dec 05 12:31:19 crc kubenswrapper[4807]: I1205 12:31:19.743459 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:31:20 crc kubenswrapper[4807]: I1205 12:31:20.169078 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:21 crc kubenswrapper[4807]: I1205 12:31:21.320893 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:21 crc kubenswrapper[4807]: I1205 12:31:21.360619 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.320709 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.339282 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.360044 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.387046 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.445186 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.479439 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.803370 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.811899 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.828217 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:26 crc kubenswrapper[4807]: I1205 12:31:26.828885 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.202484 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.203165 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-central-agent" containerID="cri-o://05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83" gracePeriod=30 Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.203212 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="proxy-httpd" containerID="cri-o://c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc" gracePeriod=30 Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.203232 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="sg-core" containerID="cri-o://f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d" gracePeriod=30 Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.203262 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-notification-agent" containerID="cri-o://6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585" gracePeriod=30 Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.253899 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.144:3000/\": read tcp 10.217.0.2:50358->10.217.0.144:3000: read: connection reset by peer" Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.826969 4807 generic.go:334] "Generic (PLEG): container finished" podID="876342fb-b348-49f2-bea7-42a4a055c3db" containerID="c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc" exitCode=0 Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.827328 4807 generic.go:334] "Generic (PLEG): container finished" podID="876342fb-b348-49f2-bea7-42a4a055c3db" containerID="f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d" exitCode=2 Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.827345 4807 generic.go:334] "Generic (PLEG): container finished" podID="876342fb-b348-49f2-bea7-42a4a055c3db" containerID="05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83" exitCode=0 Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.827048 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerDied","Data":"c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc"} Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.827388 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerDied","Data":"f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d"} Dec 05 12:31:29 crc kubenswrapper[4807]: I1205 12:31:29.827610 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerDied","Data":"05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83"} Dec 05 12:31:31 crc kubenswrapper[4807]: I1205 12:31:31.215196 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:31 crc kubenswrapper[4807]: I1205 12:31:31.215848 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-kuttl-api-log" containerID="cri-o://bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045" gracePeriod=30 Dec 05 12:31:31 crc kubenswrapper[4807]: I1205 12:31:31.215923 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-api" containerID="cri-o://d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd" gracePeriod=30 Dec 05 12:31:31 crc kubenswrapper[4807]: I1205 12:31:31.847733 4807 generic.go:334] "Generic (PLEG): container finished" podID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerID="bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045" exitCode=143 Dec 05 12:31:31 crc kubenswrapper[4807]: I1205 12:31:31.847777 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"4d510c3b-7662-4467-95d9-e28afb3d3e75","Type":"ContainerDied","Data":"bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045"} Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.107731 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.146:9322/\": read tcp 10.217.0.2:41272->10.217.0.146:9322: read: connection reset by peer" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.107769 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"https://10.217.0.146:9322/\": read tcp 10.217.0.2:41270->10.217.0.146:9322: read: connection reset by peer" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.502270 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.577353 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdmpr\" (UniqueName: \"kubernetes.io/projected/4d510c3b-7662-4467-95d9-e28afb3d3e75-kube-api-access-wdmpr\") pod \"4d510c3b-7662-4467-95d9-e28afb3d3e75\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.577421 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d510c3b-7662-4467-95d9-e28afb3d3e75-logs\") pod \"4d510c3b-7662-4467-95d9-e28afb3d3e75\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.577500 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-public-tls-certs\") pod \"4d510c3b-7662-4467-95d9-e28afb3d3e75\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.577640 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-config-data\") pod \"4d510c3b-7662-4467-95d9-e28afb3d3e75\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.577687 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-combined-ca-bundle\") pod \"4d510c3b-7662-4467-95d9-e28afb3d3e75\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.577714 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-custom-prometheus-ca\") pod \"4d510c3b-7662-4467-95d9-e28afb3d3e75\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.577774 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-internal-tls-certs\") pod \"4d510c3b-7662-4467-95d9-e28afb3d3e75\" (UID: \"4d510c3b-7662-4467-95d9-e28afb3d3e75\") " Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.578190 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d510c3b-7662-4467-95d9-e28afb3d3e75-logs" (OuterVolumeSpecName: "logs") pod "4d510c3b-7662-4467-95d9-e28afb3d3e75" (UID: "4d510c3b-7662-4467-95d9-e28afb3d3e75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.584624 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d510c3b-7662-4467-95d9-e28afb3d3e75-kube-api-access-wdmpr" (OuterVolumeSpecName: "kube-api-access-wdmpr") pod "4d510c3b-7662-4467-95d9-e28afb3d3e75" (UID: "4d510c3b-7662-4467-95d9-e28afb3d3e75"). InnerVolumeSpecName "kube-api-access-wdmpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.626360 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d510c3b-7662-4467-95d9-e28afb3d3e75" (UID: "4d510c3b-7662-4467-95d9-e28afb3d3e75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.638584 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-config-data" (OuterVolumeSpecName: "config-data") pod "4d510c3b-7662-4467-95d9-e28afb3d3e75" (UID: "4d510c3b-7662-4467-95d9-e28afb3d3e75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.641756 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "4d510c3b-7662-4467-95d9-e28afb3d3e75" (UID: "4d510c3b-7662-4467-95d9-e28afb3d3e75"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.669117 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4d510c3b-7662-4467-95d9-e28afb3d3e75" (UID: "4d510c3b-7662-4467-95d9-e28afb3d3e75"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.671026 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4d510c3b-7662-4467-95d9-e28afb3d3e75" (UID: "4d510c3b-7662-4467-95d9-e28afb3d3e75"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.685900 4807 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.686220 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdmpr\" (UniqueName: \"kubernetes.io/projected/4d510c3b-7662-4467-95d9-e28afb3d3e75-kube-api-access-wdmpr\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.686243 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d510c3b-7662-4467-95d9-e28afb3d3e75-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.686255 4807 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.686266 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.686276 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.686285 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4d510c3b-7662-4467-95d9-e28afb3d3e75-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.856331 4807 generic.go:334] "Generic (PLEG): container finished" podID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerID="d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd" exitCode=0 Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.856391 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.856389 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"4d510c3b-7662-4467-95d9-e28afb3d3e75","Type":"ContainerDied","Data":"d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd"} Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.856452 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"4d510c3b-7662-4467-95d9-e28afb3d3e75","Type":"ContainerDied","Data":"928adceeb8f13ba23a59db0a1de0f8a67339747e1a8fb529ee96a7c1309cac19"} Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.856471 4807 scope.go:117] "RemoveContainer" containerID="d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.879890 4807 scope.go:117] "RemoveContainer" containerID="bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.886911 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.895117 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.898921 4807 scope.go:117] "RemoveContainer" containerID="d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd" Dec 05 12:31:32 crc kubenswrapper[4807]: E1205 12:31:32.899371 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd\": container with ID starting with d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd not found: ID does not exist" containerID="d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.899404 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd"} err="failed to get container status \"d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd\": rpc error: code = NotFound desc = could not find container \"d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd\": container with ID starting with d5c35d736c98d6d6e80b9d244fb39b63142d13ba1e3a2f04a6fccb24120c2bbd not found: ID does not exist" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.899430 4807 scope.go:117] "RemoveContainer" containerID="bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045" Dec 05 12:31:32 crc kubenswrapper[4807]: E1205 12:31:32.899642 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045\": container with ID starting with bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045 not found: ID does not exist" containerID="bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.899661 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045"} err="failed to get container status \"bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045\": rpc error: code = NotFound desc = could not find container \"bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045\": container with ID starting with bd85d66e0a5b4984a1d5d9e3df618c6fba93e0bd269566033b5e1d7de3a4a045 not found: ID does not exist" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.910552 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:32 crc kubenswrapper[4807]: E1205 12:31:32.910963 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-kuttl-api-log" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.910993 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-kuttl-api-log" Dec 05 12:31:32 crc kubenswrapper[4807]: E1205 12:31:32.911006 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-api" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.911013 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-api" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.911275 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-kuttl-api-log" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.911309 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" containerName="watcher-api" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.912408 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.915329 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.915635 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.921961 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.926544 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.996150 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.996204 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b8308e-1108-433b-9fa8-0ae0c5528b16-logs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.996260 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.996288 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.996327 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.996432 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg72h\" (UniqueName: \"kubernetes.io/projected/95b8308e-1108-433b-9fa8-0ae0c5528b16-kube-api-access-mg72h\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:32 crc kubenswrapper[4807]: I1205 12:31:32.996613 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098049 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098100 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098130 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098163 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg72h\" (UniqueName: \"kubernetes.io/projected/95b8308e-1108-433b-9fa8-0ae0c5528b16-kube-api-access-mg72h\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098212 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098243 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098274 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b8308e-1108-433b-9fa8-0ae0c5528b16-logs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.098963 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b8308e-1108-433b-9fa8-0ae0c5528b16-logs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.102684 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.102710 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.103273 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.103813 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.105089 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.119706 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg72h\" (UniqueName: \"kubernetes.io/projected/95b8308e-1108-433b-9fa8-0ae0c5528b16-kube-api-access-mg72h\") pod \"watcher-kuttl-api-0\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.231122 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.252281 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d510c3b-7662-4467-95d9-e28afb3d3e75" path="/var/lib/kubelet/pods/4d510c3b-7662-4467-95d9-e28afb3d3e75/volumes" Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.666482 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:33 crc kubenswrapper[4807]: W1205 12:31:33.669007 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95b8308e_1108_433b_9fa8_0ae0c5528b16.slice/crio-c6618874caf0ddff1df10c82349e2ec2067a806ed3419e2e590d09e85315096c WatchSource:0}: Error finding container c6618874caf0ddff1df10c82349e2ec2067a806ed3419e2e590d09e85315096c: Status 404 returned error can't find the container with id c6618874caf0ddff1df10c82349e2ec2067a806ed3419e2e590d09e85315096c Dec 05 12:31:33 crc kubenswrapper[4807]: I1205 12:31:33.866357 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"95b8308e-1108-433b-9fa8-0ae0c5528b16","Type":"ContainerStarted","Data":"c6618874caf0ddff1df10c82349e2ec2067a806ed3419e2e590d09e85315096c"} Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.619284 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.720851 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-combined-ca-bundle\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.720910 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-config-data\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.720966 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-log-httpd\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.720985 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-ceilometer-tls-certs\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.721024 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-scripts\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.721084 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-run-httpd\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.721123 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qwwt\" (UniqueName: \"kubernetes.io/projected/876342fb-b348-49f2-bea7-42a4a055c3db-kube-api-access-4qwwt\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.721183 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-sg-core-conf-yaml\") pod \"876342fb-b348-49f2-bea7-42a4a055c3db\" (UID: \"876342fb-b348-49f2-bea7-42a4a055c3db\") " Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.722330 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.724132 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.738297 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-scripts" (OuterVolumeSpecName: "scripts") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.741216 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/876342fb-b348-49f2-bea7-42a4a055c3db-kube-api-access-4qwwt" (OuterVolumeSpecName: "kube-api-access-4qwwt") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "kube-api-access-4qwwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.747366 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.769795 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.799005 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.823102 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.823141 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qwwt\" (UniqueName: \"kubernetes.io/projected/876342fb-b348-49f2-bea7-42a4a055c3db-kube-api-access-4qwwt\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.823152 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.823160 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.823168 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/876342fb-b348-49f2-bea7-42a4a055c3db-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.823176 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.823183 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.841061 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-config-data" (OuterVolumeSpecName: "config-data") pod "876342fb-b348-49f2-bea7-42a4a055c3db" (UID: "876342fb-b348-49f2-bea7-42a4a055c3db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.883622 4807 generic.go:334] "Generic (PLEG): container finished" podID="876342fb-b348-49f2-bea7-42a4a055c3db" containerID="6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585" exitCode=0 Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.883697 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerDied","Data":"6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585"} Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.883733 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"876342fb-b348-49f2-bea7-42a4a055c3db","Type":"ContainerDied","Data":"114a6106604e71bd5f8ffa8ddf43c53dede54c755f202a83f80c6849ede2143d"} Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.883757 4807 scope.go:117] "RemoveContainer" containerID="c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.883950 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.891916 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"95b8308e-1108-433b-9fa8-0ae0c5528b16","Type":"ContainerStarted","Data":"5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01"} Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.891960 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"95b8308e-1108-433b-9fa8-0ae0c5528b16","Type":"ContainerStarted","Data":"4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d"} Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.892446 4807 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="watcher-kuttl-default/watcher-kuttl-api-0" secret="" err="secret \"watcher-watcher-kuttl-dockercfg-rmwpm\" not found" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.893300 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.919038 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz"] Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.925997 4807 scope.go:117] "RemoveContainer" containerID="f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.927040 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876342fb-b348-49f2-bea7-42a4a055c3db-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.968027 4807 scope.go:117] "RemoveContainer" containerID="6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585" Dec 05 12:31:34 crc kubenswrapper[4807]: I1205 12:31:34.968857 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-mlbzz"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.022823 4807 scope.go:117] "RemoveContainer" containerID="05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.054843 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=3.054820943 podStartE2EDuration="3.054820943s" podCreationTimestamp="2025-12-05 12:31:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:35.018454688 +0000 UTC m=+1524.512317957" watchObservedRunningTime="2025-12-05 12:31:35.054820943 +0000 UTC m=+1524.548684212" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.064863 4807 scope.go:117] "RemoveContainer" containerID="c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.067694 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc\": container with ID starting with c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc not found: ID does not exist" containerID="c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.067747 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc"} err="failed to get container status \"c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc\": rpc error: code = NotFound desc = could not find container \"c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc\": container with ID starting with c0dc0e646db14d431cd166fc086e5ec920b58c09170f53fa778ff7100722bbdc not found: ID does not exist" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.067776 4807 scope.go:117] "RemoveContainer" containerID="f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.071656 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d\": container with ID starting with f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d not found: ID does not exist" containerID="f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.071702 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d"} err="failed to get container status \"f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d\": rpc error: code = NotFound desc = could not find container \"f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d\": container with ID starting with f3d2a26edcc5edef058eebd456c87ace6945bd4330dbc11bd2481572f7253d7d not found: ID does not exist" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.071727 4807 scope.go:117] "RemoveContainer" containerID="6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.092965 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585\": container with ID starting with 6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585 not found: ID does not exist" containerID="6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.093011 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585"} err="failed to get container status \"6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585\": rpc error: code = NotFound desc = could not find container \"6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585\": container with ID starting with 6982b250af2494957765999bf0c6faeb9c38d4bcc535a280dfd1e89090b61585 not found: ID does not exist" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.093035 4807 scope.go:117] "RemoveContainer" containerID="05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.102700 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83\": container with ID starting with 05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83 not found: ID does not exist" containerID="05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.102762 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83"} err="failed to get container status \"05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83\": rpc error: code = NotFound desc = could not find container \"05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83\": container with ID starting with 05416cc4c599f1baca79b127a05ce28e3f501a24f19f626e29567d75a0137e83 not found: ID does not exist" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.151613 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcherb626-account-delete-lwq5s"] Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.152086 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="proxy-httpd" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152111 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="proxy-httpd" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.152126 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-notification-agent" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152134 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-notification-agent" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.152148 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-central-agent" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152159 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-central-agent" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.152181 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="sg-core" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152190 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="sg-core" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152398 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-notification-agent" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152425 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="ceilometer-central-agent" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152451 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="proxy-httpd" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.152465 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" containerName="sg-core" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.153295 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.185658 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherb626-account-delete-lwq5s"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.195612 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.195866 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="d535f51e-aece-4220-a465-650d815d6a12" containerName="watcher-applier" containerID="cri-o://94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b" gracePeriod=30 Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.256469 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-operator-scripts\") pod \"watcherb626-account-delete-lwq5s\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.256678 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srjwj\" (UniqueName: \"kubernetes.io/projected/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-kube-api-access-srjwj\") pod \"watcherb626-account-delete-lwq5s\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.276108 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b5ee11-c239-4de9-8ac6-0750131b4927" path="/var/lib/kubelet/pods/66b5ee11-c239-4de9-8ac6-0750131b4927/volumes" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.285828 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.285893 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.285916 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.286146 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="91b04dec-7037-469e-a32f-26aa6ef58d81" containerName="watcher-decision-engine" containerID="cri-o://0369ec80be57c6cb3ec1d477a47962495b62f837262bf38d299f2569f826322c" gracePeriod=30 Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.295815 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.314590 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.317019 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.322610 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.331545 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.331729 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.331847 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.364413 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srjwj\" (UniqueName: \"kubernetes.io/projected/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-kube-api-access-srjwj\") pod \"watcherb626-account-delete-lwq5s\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.364508 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-operator-scripts\") pod \"watcherb626-account-delete-lwq5s\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.365484 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-operator-scripts\") pod \"watcherb626-account-delete-lwq5s\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.402434 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srjwj\" (UniqueName: \"kubernetes.io/projected/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-kube-api-access-srjwj\") pod \"watcherb626-account-delete-lwq5s\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.467776 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.467826 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-scripts\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.467859 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-config-data\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.467883 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-run-httpd\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.467957 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qswxm\" (UniqueName: \"kubernetes.io/projected/5492a8b6-b2e2-421f-aa62-f777da3a0826-kube-api-access-qswxm\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.468006 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-log-httpd\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.468061 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.468108 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: E1205 12:31:35.502169 4807 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod876342fb_b348_49f2_bea7_42a4a055c3db.slice/crio-114a6106604e71bd5f8ffa8ddf43c53dede54c755f202a83f80c6849ede2143d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod876342fb_b348_49f2_bea7_42a4a055c3db.slice\": RecentStats: unable to find data in memory cache]" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.507880 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.570968 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.571042 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.571063 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-scripts\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.571090 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-config-data\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.571110 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-run-httpd\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.571166 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qswxm\" (UniqueName: \"kubernetes.io/projected/5492a8b6-b2e2-421f-aa62-f777da3a0826-kube-api-access-qswxm\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.571202 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-log-httpd\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.571247 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.576742 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-run-httpd\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.577695 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.578071 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-log-httpd\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.582862 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-config-data\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.585513 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.587240 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.588498 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-scripts\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.605412 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qswxm\" (UniqueName: \"kubernetes.io/projected/5492a8b6-b2e2-421f-aa62-f777da3a0826-kube-api-access-qswxm\") pod \"ceilometer-0\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.661147 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:35 crc kubenswrapper[4807]: I1205 12:31:35.928416 4807 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="watcher-kuttl-default/watcher-kuttl-api-0" secret="" err="secret \"watcher-watcher-kuttl-dockercfg-rmwpm\" not found" Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.055421 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherb626-account-delete-lwq5s"] Dec 05 12:31:36 crc kubenswrapper[4807]: W1205 12:31:36.070683 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e1e83a4_336b_47e0_8c84_aeaad652dcb6.slice/crio-9afc9255a8bda414052603aedced35b15aad002f8cbc56c0fb324ac7c2802613 WatchSource:0}: Error finding container 9afc9255a8bda414052603aedced35b15aad002f8cbc56c0fb324ac7c2802613: Status 404 returned error can't find the container with id 9afc9255a8bda414052603aedced35b15aad002f8cbc56c0fb324ac7c2802613 Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.082198 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.082274 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data podName:95b8308e-1108-433b-9fa8-0ae0c5528b16 nodeName:}" failed. No retries permitted until 2025-12-05 12:31:36.582250086 +0000 UTC m=+1526.076113365 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data") pod "watcher-kuttl-api-0" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16") : secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:36 crc kubenswrapper[4807]: W1205 12:31:36.254241 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5492a8b6_b2e2_421f_aa62_f777da3a0826.slice/crio-3ee3fbaa0a64fab92fbb058f2d08d6256ede77a785c423504565fe34c45609da WatchSource:0}: Error finding container 3ee3fbaa0a64fab92fbb058f2d08d6256ede77a785c423504565fe34c45609da: Status 404 returned error can't find the container with id 3ee3fbaa0a64fab92fbb058f2d08d6256ede77a785c423504565fe34c45609da Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.257818 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.363079 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.366252 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.368555 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.368614 4807 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="d535f51e-aece-4220-a465-650d815d6a12" containerName="watcher-applier" Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.591586 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:36 crc kubenswrapper[4807]: E1205 12:31:36.591929 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data podName:95b8308e-1108-433b-9fa8-0ae0c5528b16 nodeName:}" failed. No retries permitted until 2025-12-05 12:31:37.591909184 +0000 UTC m=+1527.085772453 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data") pod "watcher-kuttl-api-0" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16") : secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.936205 4807 generic.go:334] "Generic (PLEG): container finished" podID="7e1e83a4-336b-47e0-8c84-aeaad652dcb6" containerID="d310361021ce548f3a27d4d0ad3409df1ad5a47e895739cd5e779a37004cee08" exitCode=0 Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.936261 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" event={"ID":"7e1e83a4-336b-47e0-8c84-aeaad652dcb6","Type":"ContainerDied","Data":"d310361021ce548f3a27d4d0ad3409df1ad5a47e895739cd5e779a37004cee08"} Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.936491 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" event={"ID":"7e1e83a4-336b-47e0-8c84-aeaad652dcb6","Type":"ContainerStarted","Data":"9afc9255a8bda414052603aedced35b15aad002f8cbc56c0fb324ac7c2802613"} Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.937486 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerStarted","Data":"3ee3fbaa0a64fab92fbb058f2d08d6256ede77a785c423504565fe34c45609da"} Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.937517 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.937687 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-kuttl-api-log" containerID="cri-o://4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d" gracePeriod=30 Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.937738 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-api" containerID="cri-o://5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01" gracePeriod=30 Dec 05 12:31:36 crc kubenswrapper[4807]: I1205 12:31:36.955637 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.149:9322/\": EOF" Dec 05 12:31:37 crc kubenswrapper[4807]: I1205 12:31:37.246864 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="876342fb-b348-49f2-bea7-42a4a055c3db" path="/var/lib/kubelet/pods/876342fb-b348-49f2-bea7-42a4a055c3db/volumes" Dec 05 12:31:37 crc kubenswrapper[4807]: E1205 12:31:37.607969 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:37 crc kubenswrapper[4807]: E1205 12:31:37.608077 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data podName:95b8308e-1108-433b-9fa8-0ae0c5528b16 nodeName:}" failed. No retries permitted until 2025-12-05 12:31:39.608057493 +0000 UTC m=+1529.101920762 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data") pod "watcher-kuttl-api-0" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16") : secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:37 crc kubenswrapper[4807]: I1205 12:31:37.947055 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerStarted","Data":"e9fc383aa48056c9d280d6c67e10fd1b2bacd80d743d6798c0d2283b9da6fc96"} Dec 05 12:31:37 crc kubenswrapper[4807]: I1205 12:31:37.947351 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerStarted","Data":"3ea6fcdbd7bea4253e5d59651132866093218d0ecc3d72135d3e95f0d805c56d"} Dec 05 12:31:37 crc kubenswrapper[4807]: I1205 12:31:37.949215 4807 generic.go:334] "Generic (PLEG): container finished" podID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerID="4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d" exitCode=143 Dec 05 12:31:37 crc kubenswrapper[4807]: I1205 12:31:37.949248 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"95b8308e-1108-433b-9fa8-0ae0c5528b16","Type":"ContainerDied","Data":"4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d"} Dec 05 12:31:37 crc kubenswrapper[4807]: I1205 12:31:37.976545 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.231762 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.450203 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.638565 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-operator-scripts\") pod \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.638716 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srjwj\" (UniqueName: \"kubernetes.io/projected/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-kube-api-access-srjwj\") pod \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\" (UID: \"7e1e83a4-336b-47e0-8c84-aeaad652dcb6\") " Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.639061 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e1e83a4-336b-47e0-8c84-aeaad652dcb6" (UID: "7e1e83a4-336b-47e0-8c84-aeaad652dcb6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.639311 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.643738 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-kube-api-access-srjwj" (OuterVolumeSpecName: "kube-api-access-srjwj") pod "7e1e83a4-336b-47e0-8c84-aeaad652dcb6" (UID: "7e1e83a4-336b-47e0-8c84-aeaad652dcb6"). InnerVolumeSpecName "kube-api-access-srjwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.741737 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srjwj\" (UniqueName: \"kubernetes.io/projected/7e1e83a4-336b-47e0-8c84-aeaad652dcb6-kube-api-access-srjwj\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.969455 4807 generic.go:334] "Generic (PLEG): container finished" podID="d535f51e-aece-4220-a465-650d815d6a12" containerID="94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b" exitCode=0 Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.969624 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"d535f51e-aece-4220-a465-650d815d6a12","Type":"ContainerDied","Data":"94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b"} Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.972894 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.973539 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherb626-account-delete-lwq5s" event={"ID":"7e1e83a4-336b-47e0-8c84-aeaad652dcb6","Type":"ContainerDied","Data":"9afc9255a8bda414052603aedced35b15aad002f8cbc56c0fb324ac7c2802613"} Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.973592 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9afc9255a8bda414052603aedced35b15aad002f8cbc56c0fb324ac7c2802613" Dec 05 12:31:38 crc kubenswrapper[4807]: I1205 12:31:38.978028 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerStarted","Data":"8a1ccb08f40648be310ccb8ef3502ec75172925aa98a1c199212a8e49d4fa2d1"} Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.082820 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.149:9322/\": read tcp 10.217.0.2:45190->10.217.0.149:9322: read: connection reset by peer" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.083282 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.149:9322/\": dial tcp 10.217.0.149:9322: connect: connection refused" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.398041 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.568754 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-combined-ca-bundle\") pod \"d535f51e-aece-4220-a465-650d815d6a12\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.568918 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-config-data\") pod \"d535f51e-aece-4220-a465-650d815d6a12\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.569472 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d535f51e-aece-4220-a465-650d815d6a12-logs\") pod \"d535f51e-aece-4220-a465-650d815d6a12\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.569503 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5vx7\" (UniqueName: \"kubernetes.io/projected/d535f51e-aece-4220-a465-650d815d6a12-kube-api-access-r5vx7\") pod \"d535f51e-aece-4220-a465-650d815d6a12\" (UID: \"d535f51e-aece-4220-a465-650d815d6a12\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.569872 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d535f51e-aece-4220-a465-650d815d6a12-logs" (OuterVolumeSpecName: "logs") pod "d535f51e-aece-4220-a465-650d815d6a12" (UID: "d535f51e-aece-4220-a465-650d815d6a12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.570278 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d535f51e-aece-4220-a465-650d815d6a12-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.577303 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d535f51e-aece-4220-a465-650d815d6a12-kube-api-access-r5vx7" (OuterVolumeSpecName: "kube-api-access-r5vx7") pod "d535f51e-aece-4220-a465-650d815d6a12" (UID: "d535f51e-aece-4220-a465-650d815d6a12"). InnerVolumeSpecName "kube-api-access-r5vx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.603900 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.613747 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d535f51e-aece-4220-a465-650d815d6a12" (UID: "d535f51e-aece-4220-a465-650d815d6a12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.645725 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-config-data" (OuterVolumeSpecName: "config-data") pod "d535f51e-aece-4220-a465-650d815d6a12" (UID: "d535f51e-aece-4220-a465-650d815d6a12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.672306 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.672346 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5vx7\" (UniqueName: \"kubernetes.io/projected/d535f51e-aece-4220-a465-650d815d6a12-kube-api-access-r5vx7\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.672362 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d535f51e-aece-4220-a465-650d815d6a12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: E1205 12:31:39.672438 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:39 crc kubenswrapper[4807]: E1205 12:31:39.672539 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data podName:95b8308e-1108-433b-9fa8-0ae0c5528b16 nodeName:}" failed. No retries permitted until 2025-12-05 12:31:43.672504442 +0000 UTC m=+1533.166367771 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data") pod "watcher-kuttl-api-0" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16") : secret "watcher-kuttl-api-config-data" not found Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.773817 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-custom-prometheus-ca\") pod \"95b8308e-1108-433b-9fa8-0ae0c5528b16\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.773940 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data\") pod \"95b8308e-1108-433b-9fa8-0ae0c5528b16\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.774019 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b8308e-1108-433b-9fa8-0ae0c5528b16-logs\") pod \"95b8308e-1108-433b-9fa8-0ae0c5528b16\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.774102 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-internal-tls-certs\") pod \"95b8308e-1108-433b-9fa8-0ae0c5528b16\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.774174 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg72h\" (UniqueName: \"kubernetes.io/projected/95b8308e-1108-433b-9fa8-0ae0c5528b16-kube-api-access-mg72h\") pod \"95b8308e-1108-433b-9fa8-0ae0c5528b16\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.774230 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-public-tls-certs\") pod \"95b8308e-1108-433b-9fa8-0ae0c5528b16\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.774261 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-combined-ca-bundle\") pod \"95b8308e-1108-433b-9fa8-0ae0c5528b16\" (UID: \"95b8308e-1108-433b-9fa8-0ae0c5528b16\") " Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.775110 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b8308e-1108-433b-9fa8-0ae0c5528b16-logs" (OuterVolumeSpecName: "logs") pod "95b8308e-1108-433b-9fa8-0ae0c5528b16" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.778236 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b8308e-1108-433b-9fa8-0ae0c5528b16-kube-api-access-mg72h" (OuterVolumeSpecName: "kube-api-access-mg72h") pod "95b8308e-1108-433b-9fa8-0ae0c5528b16" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16"). InnerVolumeSpecName "kube-api-access-mg72h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.811552 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95b8308e-1108-433b-9fa8-0ae0c5528b16" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.832759 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "95b8308e-1108-433b-9fa8-0ae0c5528b16" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.834646 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "95b8308e-1108-433b-9fa8-0ae0c5528b16" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.843278 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "95b8308e-1108-433b-9fa8-0ae0c5528b16" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.852647 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data" (OuterVolumeSpecName: "config-data") pod "95b8308e-1108-433b-9fa8-0ae0c5528b16" (UID: "95b8308e-1108-433b-9fa8-0ae0c5528b16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.876471 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg72h\" (UniqueName: \"kubernetes.io/projected/95b8308e-1108-433b-9fa8-0ae0c5528b16-kube-api-access-mg72h\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.876508 4807 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.876539 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.876552 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.876568 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.876579 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b8308e-1108-433b-9fa8-0ae0c5528b16-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:39 crc kubenswrapper[4807]: I1205 12:31:39.876589 4807 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b8308e-1108-433b-9fa8-0ae0c5528b16-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.012066 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcherb626-account-delete-lwq5s"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.026469 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-b626-account-create-update-5v8wr"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.028470 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.028662 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"d535f51e-aece-4220-a465-650d815d6a12","Type":"ContainerDied","Data":"79633ae00afbcde37575d17407bae0dd3405e33f624232bb4a604edd46888946"} Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.028732 4807 scope.go:117] "RemoveContainer" containerID="94f7e2bb07385fa0e8c73cb9a2c71b05c4b444891fc5cc4187647122d7436d4b" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.038572 4807 generic.go:334] "Generic (PLEG): container finished" podID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerID="5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01" exitCode=0 Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.038618 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.038675 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"95b8308e-1108-433b-9fa8-0ae0c5528b16","Type":"ContainerDied","Data":"5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01"} Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.038742 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"95b8308e-1108-433b-9fa8-0ae0c5528b16","Type":"ContainerDied","Data":"c6618874caf0ddff1df10c82349e2ec2067a806ed3419e2e590d09e85315096c"} Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.041475 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lpj9k"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.046667 4807 generic.go:334] "Generic (PLEG): container finished" podID="91b04dec-7037-469e-a32f-26aa6ef58d81" containerID="0369ec80be57c6cb3ec1d477a47962495b62f837262bf38d299f2569f826322c" exitCode=0 Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.046703 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"91b04dec-7037-469e-a32f-26aa6ef58d81","Type":"ContainerDied","Data":"0369ec80be57c6cb3ec1d477a47962495b62f837262bf38d299f2569f826322c"} Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.054729 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcherb626-account-delete-lwq5s"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.099657 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-b626-account-create-update-5v8wr"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.106132 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lpj9k"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.134778 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.140755 4807 scope.go:117] "RemoveContainer" containerID="5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.144201 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.154194 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.165012 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.188293 4807 scope.go:117] "RemoveContainer" containerID="4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.213807 4807 scope.go:117] "RemoveContainer" containerID="5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01" Dec 05 12:31:40 crc kubenswrapper[4807]: E1205 12:31:40.215438 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01\": container with ID starting with 5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01 not found: ID does not exist" containerID="5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.215472 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01"} err="failed to get container status \"5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01\": rpc error: code = NotFound desc = could not find container \"5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01\": container with ID starting with 5dde05bb02e0f57661326468a87aa51ce6b9860e9927a7de090604e93cc6bf01 not found: ID does not exist" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.215496 4807 scope.go:117] "RemoveContainer" containerID="4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.216825 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:40 crc kubenswrapper[4807]: E1205 12:31:40.217170 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d\": container with ID starting with 4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d not found: ID does not exist" containerID="4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.217197 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d"} err="failed to get container status \"4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d\": rpc error: code = NotFound desc = could not find container \"4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d\": container with ID starting with 4a0ab5aca0315ef5b39a79c03bbdd468c3da5aaa504d7efade89d422af579b0d not found: ID does not exist" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.389091 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-combined-ca-bundle\") pod \"91b04dec-7037-469e-a32f-26aa6ef58d81\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.389179 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-custom-prometheus-ca\") pod \"91b04dec-7037-469e-a32f-26aa6ef58d81\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.389376 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-config-data\") pod \"91b04dec-7037-469e-a32f-26aa6ef58d81\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.389413 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp2ht\" (UniqueName: \"kubernetes.io/projected/91b04dec-7037-469e-a32f-26aa6ef58d81-kube-api-access-tp2ht\") pod \"91b04dec-7037-469e-a32f-26aa6ef58d81\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.389464 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91b04dec-7037-469e-a32f-26aa6ef58d81-logs\") pod \"91b04dec-7037-469e-a32f-26aa6ef58d81\" (UID: \"91b04dec-7037-469e-a32f-26aa6ef58d81\") " Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.391932 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91b04dec-7037-469e-a32f-26aa6ef58d81-logs" (OuterVolumeSpecName: "logs") pod "91b04dec-7037-469e-a32f-26aa6ef58d81" (UID: "91b04dec-7037-469e-a32f-26aa6ef58d81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.394482 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91b04dec-7037-469e-a32f-26aa6ef58d81-kube-api-access-tp2ht" (OuterVolumeSpecName: "kube-api-access-tp2ht") pod "91b04dec-7037-469e-a32f-26aa6ef58d81" (UID: "91b04dec-7037-469e-a32f-26aa6ef58d81"). InnerVolumeSpecName "kube-api-access-tp2ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.415609 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91b04dec-7037-469e-a32f-26aa6ef58d81" (UID: "91b04dec-7037-469e-a32f-26aa6ef58d81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.428972 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "91b04dec-7037-469e-a32f-26aa6ef58d81" (UID: "91b04dec-7037-469e-a32f-26aa6ef58d81"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.445586 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-config-data" (OuterVolumeSpecName: "config-data") pod "91b04dec-7037-469e-a32f-26aa6ef58d81" (UID: "91b04dec-7037-469e-a32f-26aa6ef58d81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.491644 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.491677 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.491686 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp2ht\" (UniqueName: \"kubernetes.io/projected/91b04dec-7037-469e-a32f-26aa6ef58d81-kube-api-access-tp2ht\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.491695 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91b04dec-7037-469e-a32f-26aa6ef58d81-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:40 crc kubenswrapper[4807]: I1205 12:31:40.491705 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91b04dec-7037-469e-a32f-26aa6ef58d81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.056842 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"91b04dec-7037-469e-a32f-26aa6ef58d81","Type":"ContainerDied","Data":"be301c896545720b73dead96128195206ed663cd75fe4a997999a9393870982a"} Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.056884 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.056950 4807 scope.go:117] "RemoveContainer" containerID="0369ec80be57c6cb3ec1d477a47962495b62f837262bf38d299f2569f826322c" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.062150 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerStarted","Data":"eeb3f986838869505a062c8b241a51624ce8e4225f841c0e619940f7de2e217a"} Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.062302 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-central-agent" containerID="cri-o://3ea6fcdbd7bea4253e5d59651132866093218d0ecc3d72135d3e95f0d805c56d" gracePeriod=30 Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.062804 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="proxy-httpd" containerID="cri-o://eeb3f986838869505a062c8b241a51624ce8e4225f841c0e619940f7de2e217a" gracePeriod=30 Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.062848 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-notification-agent" containerID="cri-o://e9fc383aa48056c9d280d6c67e10fd1b2bacd80d743d6798c0d2283b9da6fc96" gracePeriod=30 Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.062975 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.063006 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="sg-core" containerID="cri-o://8a1ccb08f40648be310ccb8ef3502ec75172925aa98a1c199212a8e49d4fa2d1" gracePeriod=30 Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.115646 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.347923225 podStartE2EDuration="6.115626948s" podCreationTimestamp="2025-12-05 12:31:35 +0000 UTC" firstStartedPulling="2025-12-05 12:31:36.256675449 +0000 UTC m=+1525.750538718" lastFinishedPulling="2025-12-05 12:31:40.024379172 +0000 UTC m=+1529.518242441" observedRunningTime="2025-12-05 12:31:41.099957267 +0000 UTC m=+1530.593820536" watchObservedRunningTime="2025-12-05 12:31:41.115626948 +0000 UTC m=+1530.609490217" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.127365 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.134101 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.246908 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c" path="/var/lib/kubelet/pods/3f65d42c-0e0b-4ed2-a6ef-45bc3dab527c/volumes" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.247500 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e1e83a4-336b-47e0-8c84-aeaad652dcb6" path="/var/lib/kubelet/pods/7e1e83a4-336b-47e0-8c84-aeaad652dcb6/volumes" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.248119 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91b04dec-7037-469e-a32f-26aa6ef58d81" path="/var/lib/kubelet/pods/91b04dec-7037-469e-a32f-26aa6ef58d81/volumes" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.249308 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" path="/var/lib/kubelet/pods/95b8308e-1108-433b-9fa8-0ae0c5528b16/volumes" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.250039 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d535f51e-aece-4220-a465-650d815d6a12" path="/var/lib/kubelet/pods/d535f51e-aece-4220-a465-650d815d6a12/volumes" Dec 05 12:31:41 crc kubenswrapper[4807]: I1205 12:31:41.250510 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe61e51f-15fe-4b89-a025-4d3949ecbbc2" path="/var/lib/kubelet/pods/fe61e51f-15fe-4b89-a025-4d3949ecbbc2/volumes" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074404 4807 generic.go:334] "Generic (PLEG): container finished" podID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerID="eeb3f986838869505a062c8b241a51624ce8e4225f841c0e619940f7de2e217a" exitCode=0 Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074809 4807 generic.go:334] "Generic (PLEG): container finished" podID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerID="8a1ccb08f40648be310ccb8ef3502ec75172925aa98a1c199212a8e49d4fa2d1" exitCode=2 Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074825 4807 generic.go:334] "Generic (PLEG): container finished" podID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerID="e9fc383aa48056c9d280d6c67e10fd1b2bacd80d743d6798c0d2283b9da6fc96" exitCode=0 Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074848 4807 generic.go:334] "Generic (PLEG): container finished" podID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerID="3ea6fcdbd7bea4253e5d59651132866093218d0ecc3d72135d3e95f0d805c56d" exitCode=0 Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074478 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerDied","Data":"eeb3f986838869505a062c8b241a51624ce8e4225f841c0e619940f7de2e217a"} Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074933 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerDied","Data":"8a1ccb08f40648be310ccb8ef3502ec75172925aa98a1c199212a8e49d4fa2d1"} Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074955 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerDied","Data":"e9fc383aa48056c9d280d6c67e10fd1b2bacd80d743d6798c0d2283b9da6fc96"} Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.074968 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerDied","Data":"3ea6fcdbd7bea4253e5d59651132866093218d0ecc3d72135d3e95f0d805c56d"} Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.339798 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-mw4nj"] Dec 05 12:31:42 crc kubenswrapper[4807]: E1205 12:31:42.341842 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-kuttl-api-log" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.341864 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-kuttl-api-log" Dec 05 12:31:42 crc kubenswrapper[4807]: E1205 12:31:42.341875 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b04dec-7037-469e-a32f-26aa6ef58d81" containerName="watcher-decision-engine" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.341882 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b04dec-7037-469e-a32f-26aa6ef58d81" containerName="watcher-decision-engine" Dec 05 12:31:42 crc kubenswrapper[4807]: E1205 12:31:42.341906 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e1e83a4-336b-47e0-8c84-aeaad652dcb6" containerName="mariadb-account-delete" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.341922 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e1e83a4-336b-47e0-8c84-aeaad652dcb6" containerName="mariadb-account-delete" Dec 05 12:31:42 crc kubenswrapper[4807]: E1205 12:31:42.341934 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-api" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.341939 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-api" Dec 05 12:31:42 crc kubenswrapper[4807]: E1205 12:31:42.341951 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d535f51e-aece-4220-a465-650d815d6a12" containerName="watcher-applier" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.341957 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d535f51e-aece-4220-a465-650d815d6a12" containerName="watcher-applier" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.342094 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d535f51e-aece-4220-a465-650d815d6a12" containerName="watcher-applier" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.342105 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="91b04dec-7037-469e-a32f-26aa6ef58d81" containerName="watcher-decision-engine" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.342117 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-kuttl-api-log" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.342129 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b8308e-1108-433b-9fa8-0ae0c5528b16" containerName="watcher-api" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.342138 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e1e83a4-336b-47e0-8c84-aeaad652dcb6" containerName="mariadb-account-delete" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.342714 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.358225 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-mw4nj"] Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.421596 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dw5z\" (UniqueName: \"kubernetes.io/projected/6a0cd735-40cf-49aa-be5a-2378115ea996-kube-api-access-8dw5z\") pod \"watcher-db-create-mw4nj\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.421658 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a0cd735-40cf-49aa-be5a-2378115ea996-operator-scripts\") pod \"watcher-db-create-mw4nj\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.445225 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2"] Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.446490 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.449310 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.457781 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2"] Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.523656 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdcls\" (UniqueName: \"kubernetes.io/projected/cc06c206-269f-48e7-b014-f1a7d9f596ac-kube-api-access-bdcls\") pod \"watcher-df4f-account-create-update-2g9k2\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.523703 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc06c206-269f-48e7-b014-f1a7d9f596ac-operator-scripts\") pod \"watcher-df4f-account-create-update-2g9k2\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.523807 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dw5z\" (UniqueName: \"kubernetes.io/projected/6a0cd735-40cf-49aa-be5a-2378115ea996-kube-api-access-8dw5z\") pod \"watcher-db-create-mw4nj\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.523840 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a0cd735-40cf-49aa-be5a-2378115ea996-operator-scripts\") pod \"watcher-db-create-mw4nj\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.524447 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a0cd735-40cf-49aa-be5a-2378115ea996-operator-scripts\") pod \"watcher-db-create-mw4nj\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.562259 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dw5z\" (UniqueName: \"kubernetes.io/projected/6a0cd735-40cf-49aa-be5a-2378115ea996-kube-api-access-8dw5z\") pod \"watcher-db-create-mw4nj\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.624996 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdcls\" (UniqueName: \"kubernetes.io/projected/cc06c206-269f-48e7-b014-f1a7d9f596ac-kube-api-access-bdcls\") pod \"watcher-df4f-account-create-update-2g9k2\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.625047 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc06c206-269f-48e7-b014-f1a7d9f596ac-operator-scripts\") pod \"watcher-df4f-account-create-update-2g9k2\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.625699 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc06c206-269f-48e7-b014-f1a7d9f596ac-operator-scripts\") pod \"watcher-df4f-account-create-update-2g9k2\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.653347 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdcls\" (UniqueName: \"kubernetes.io/projected/cc06c206-269f-48e7-b014-f1a7d9f596ac-kube-api-access-bdcls\") pod \"watcher-df4f-account-create-update-2g9k2\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.660442 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.776943 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.789095 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.827252 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-ceilometer-tls-certs\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.827913 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qswxm\" (UniqueName: \"kubernetes.io/projected/5492a8b6-b2e2-421f-aa62-f777da3a0826-kube-api-access-qswxm\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.827962 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-log-httpd\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.827999 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-sg-core-conf-yaml\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.828043 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-scripts\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.828081 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-run-httpd\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.828130 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-combined-ca-bundle\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.828198 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-config-data\") pod \"5492a8b6-b2e2-421f-aa62-f777da3a0826\" (UID: \"5492a8b6-b2e2-421f-aa62-f777da3a0826\") " Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.835125 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.835826 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.841549 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-scripts" (OuterVolumeSpecName: "scripts") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.842234 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5492a8b6-b2e2-421f-aa62-f777da3a0826-kube-api-access-qswxm" (OuterVolumeSpecName: "kube-api-access-qswxm") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "kube-api-access-qswxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.867420 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.915650 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.930161 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.930231 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qswxm\" (UniqueName: \"kubernetes.io/projected/5492a8b6-b2e2-421f-aa62-f777da3a0826-kube-api-access-qswxm\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.930247 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.930259 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.930270 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.930281 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5492a8b6-b2e2-421f-aa62-f777da3a0826-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.991610 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-config-data" (OuterVolumeSpecName: "config-data") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:42 crc kubenswrapper[4807]: I1205 12:31:42.991928 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5492a8b6-b2e2-421f-aa62-f777da3a0826" (UID: "5492a8b6-b2e2-421f-aa62-f777da3a0826"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.031362 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.031388 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5492a8b6-b2e2-421f-aa62-f777da3a0826-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.086008 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5492a8b6-b2e2-421f-aa62-f777da3a0826","Type":"ContainerDied","Data":"3ee3fbaa0a64fab92fbb058f2d08d6256ede77a785c423504565fe34c45609da"} Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.086056 4807 scope.go:117] "RemoveContainer" containerID="eeb3f986838869505a062c8b241a51624ce8e4225f841c0e619940f7de2e217a" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.086155 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.115637 4807 scope.go:117] "RemoveContainer" containerID="8a1ccb08f40648be310ccb8ef3502ec75172925aa98a1c199212a8e49d4fa2d1" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.125466 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.146348 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162330 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:43 crc kubenswrapper[4807]: E1205 12:31:43.162704 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-notification-agent" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162721 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-notification-agent" Dec 05 12:31:43 crc kubenswrapper[4807]: E1205 12:31:43.162737 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="sg-core" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162742 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="sg-core" Dec 05 12:31:43 crc kubenswrapper[4807]: E1205 12:31:43.162760 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="proxy-httpd" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162766 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="proxy-httpd" Dec 05 12:31:43 crc kubenswrapper[4807]: E1205 12:31:43.162787 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-central-agent" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162792 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-central-agent" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162942 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="proxy-httpd" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162956 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="sg-core" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162971 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-notification-agent" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.162983 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" containerName="ceilometer-central-agent" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.170936 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.171135 4807 scope.go:117] "RemoveContainer" containerID="e9fc383aa48056c9d280d6c67e10fd1b2bacd80d743d6798c0d2283b9da6fc96" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.174726 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.174777 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.178370 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.198644 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.233115 4807 scope.go:117] "RemoveContainer" containerID="3ea6fcdbd7bea4253e5d59651132866093218d0ecc3d72135d3e95f0d805c56d" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234224 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qggsm\" (UniqueName: \"kubernetes.io/projected/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-kube-api-access-qggsm\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234290 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-log-httpd\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234334 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234354 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-config-data\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234381 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234408 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-scripts\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234434 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-run-httpd\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.234452 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.271855 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5492a8b6-b2e2-421f-aa62-f777da3a0826" path="/var/lib/kubelet/pods/5492a8b6-b2e2-421f-aa62-f777da3a0826/volumes" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.273227 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-mw4nj"] Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.337504 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-scripts\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.337894 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-run-httpd\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.337928 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.337952 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qggsm\" (UniqueName: \"kubernetes.io/projected/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-kube-api-access-qggsm\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.338058 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-log-httpd\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.338130 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.338156 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-config-data\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.338188 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.338442 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-run-httpd\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.338747 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-log-httpd\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.345787 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.346334 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.350461 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-config-data\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.350564 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.351468 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-scripts\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.352138 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2"] Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.358870 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qggsm\" (UniqueName: \"kubernetes.io/projected/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-kube-api-access-qggsm\") pod \"ceilometer-0\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:43 crc kubenswrapper[4807]: I1205 12:31:43.551004 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.003124 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:31:44 crc kubenswrapper[4807]: W1205 12:31:44.015199 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd39b4a0_3725_4ae9_bad5_8cf81a720a33.slice/crio-d72e7291d30b9f78288122d5c660d088052512c1c1bb67ed827b91412c84ef9b WatchSource:0}: Error finding container d72e7291d30b9f78288122d5c660d088052512c1c1bb67ed827b91412c84ef9b: Status 404 returned error can't find the container with id d72e7291d30b9f78288122d5c660d088052512c1c1bb67ed827b91412c84ef9b Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.095983 4807 generic.go:334] "Generic (PLEG): container finished" podID="cc06c206-269f-48e7-b014-f1a7d9f596ac" containerID="eb2aafa140e8b2537becc06f69ead552cf44fe114bf70966e0de9a32ffa41f73" exitCode=0 Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.096095 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" event={"ID":"cc06c206-269f-48e7-b014-f1a7d9f596ac","Type":"ContainerDied","Data":"eb2aafa140e8b2537becc06f69ead552cf44fe114bf70966e0de9a32ffa41f73"} Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.096173 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" event={"ID":"cc06c206-269f-48e7-b014-f1a7d9f596ac","Type":"ContainerStarted","Data":"fd87b62f282559d58b0acbea3c67179df0d4cc20fa3c3ac424719795c7dbaa88"} Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.097870 4807 generic.go:334] "Generic (PLEG): container finished" podID="6a0cd735-40cf-49aa-be5a-2378115ea996" containerID="cb9d46c5057b570f2d7178d833eac2de5e7af8162e538e28de8c43d818fb4866" exitCode=0 Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.097970 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-mw4nj" event={"ID":"6a0cd735-40cf-49aa-be5a-2378115ea996","Type":"ContainerDied","Data":"cb9d46c5057b570f2d7178d833eac2de5e7af8162e538e28de8c43d818fb4866"} Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.098004 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-mw4nj" event={"ID":"6a0cd735-40cf-49aa-be5a-2378115ea996","Type":"ContainerStarted","Data":"0944579106bda5e2acf54741286bff810342be45f766e8919c557bf791b119e2"} Dec 05 12:31:44 crc kubenswrapper[4807]: I1205 12:31:44.101776 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerStarted","Data":"d72e7291d30b9f78288122d5c660d088052512c1c1bb67ed827b91412c84ef9b"} Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.109505 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerStarted","Data":"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac"} Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.574443 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.644265 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.676292 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc06c206-269f-48e7-b014-f1a7d9f596ac-operator-scripts\") pod \"cc06c206-269f-48e7-b014-f1a7d9f596ac\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.676413 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdcls\" (UniqueName: \"kubernetes.io/projected/cc06c206-269f-48e7-b014-f1a7d9f596ac-kube-api-access-bdcls\") pod \"cc06c206-269f-48e7-b014-f1a7d9f596ac\" (UID: \"cc06c206-269f-48e7-b014-f1a7d9f596ac\") " Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.676448 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a0cd735-40cf-49aa-be5a-2378115ea996-operator-scripts\") pod \"6a0cd735-40cf-49aa-be5a-2378115ea996\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.676508 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dw5z\" (UniqueName: \"kubernetes.io/projected/6a0cd735-40cf-49aa-be5a-2378115ea996-kube-api-access-8dw5z\") pod \"6a0cd735-40cf-49aa-be5a-2378115ea996\" (UID: \"6a0cd735-40cf-49aa-be5a-2378115ea996\") " Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.678331 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc06c206-269f-48e7-b014-f1a7d9f596ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc06c206-269f-48e7-b014-f1a7d9f596ac" (UID: "cc06c206-269f-48e7-b014-f1a7d9f596ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.679442 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a0cd735-40cf-49aa-be5a-2378115ea996-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a0cd735-40cf-49aa-be5a-2378115ea996" (UID: "6a0cd735-40cf-49aa-be5a-2378115ea996"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.683677 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0cd735-40cf-49aa-be5a-2378115ea996-kube-api-access-8dw5z" (OuterVolumeSpecName: "kube-api-access-8dw5z") pod "6a0cd735-40cf-49aa-be5a-2378115ea996" (UID: "6a0cd735-40cf-49aa-be5a-2378115ea996"). InnerVolumeSpecName "kube-api-access-8dw5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.696773 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc06c206-269f-48e7-b014-f1a7d9f596ac-kube-api-access-bdcls" (OuterVolumeSpecName: "kube-api-access-bdcls") pod "cc06c206-269f-48e7-b014-f1a7d9f596ac" (UID: "cc06c206-269f-48e7-b014-f1a7d9f596ac"). InnerVolumeSpecName "kube-api-access-bdcls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.778433 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc06c206-269f-48e7-b014-f1a7d9f596ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.778477 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdcls\" (UniqueName: \"kubernetes.io/projected/cc06c206-269f-48e7-b014-f1a7d9f596ac-kube-api-access-bdcls\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.778491 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a0cd735-40cf-49aa-be5a-2378115ea996-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:45 crc kubenswrapper[4807]: I1205 12:31:45.778503 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dw5z\" (UniqueName: \"kubernetes.io/projected/6a0cd735-40cf-49aa-be5a-2378115ea996-kube-api-access-8dw5z\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:46 crc kubenswrapper[4807]: I1205 12:31:46.122660 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerStarted","Data":"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5"} Dec 05 12:31:46 crc kubenswrapper[4807]: I1205 12:31:46.124885 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" event={"ID":"cc06c206-269f-48e7-b014-f1a7d9f596ac","Type":"ContainerDied","Data":"fd87b62f282559d58b0acbea3c67179df0d4cc20fa3c3ac424719795c7dbaa88"} Dec 05 12:31:46 crc kubenswrapper[4807]: I1205 12:31:46.124928 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd87b62f282559d58b0acbea3c67179df0d4cc20fa3c3ac424719795c7dbaa88" Dec 05 12:31:46 crc kubenswrapper[4807]: I1205 12:31:46.125003 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2" Dec 05 12:31:46 crc kubenswrapper[4807]: I1205 12:31:46.129770 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-mw4nj" event={"ID":"6a0cd735-40cf-49aa-be5a-2378115ea996","Type":"ContainerDied","Data":"0944579106bda5e2acf54741286bff810342be45f766e8919c557bf791b119e2"} Dec 05 12:31:46 crc kubenswrapper[4807]: I1205 12:31:46.129815 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0944579106bda5e2acf54741286bff810342be45f766e8919c557bf791b119e2" Dec 05 12:31:46 crc kubenswrapper[4807]: I1205 12:31:46.130425 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-mw4nj" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.140307 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerStarted","Data":"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9"} Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.765882 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh"] Dec 05 12:31:47 crc kubenswrapper[4807]: E1205 12:31:47.766869 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc06c206-269f-48e7-b014-f1a7d9f596ac" containerName="mariadb-account-create-update" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.766973 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc06c206-269f-48e7-b014-f1a7d9f596ac" containerName="mariadb-account-create-update" Dec 05 12:31:47 crc kubenswrapper[4807]: E1205 12:31:47.767049 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0cd735-40cf-49aa-be5a-2378115ea996" containerName="mariadb-database-create" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.767125 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0cd735-40cf-49aa-be5a-2378115ea996" containerName="mariadb-database-create" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.767372 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc06c206-269f-48e7-b014-f1a7d9f596ac" containerName="mariadb-account-create-update" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.773408 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0cd735-40cf-49aa-be5a-2378115ea996" containerName="mariadb-database-create" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.774251 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.775669 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh"] Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.780774 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.781898 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-9zr5r" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.838720 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgdpc\" (UniqueName: \"kubernetes.io/projected/c6c66668-1556-427e-be1d-1d36a93e1369-kube-api-access-rgdpc\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.838788 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-db-sync-config-data\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.839001 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-config-data\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.839128 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.941013 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-config-data\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.941112 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.941202 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgdpc\" (UniqueName: \"kubernetes.io/projected/c6c66668-1556-427e-be1d-1d36a93e1369-kube-api-access-rgdpc\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.941240 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-db-sync-config-data\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.948655 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-db-sync-config-data\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.948967 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-config-data\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.949035 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:47 crc kubenswrapper[4807]: I1205 12:31:47.965277 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgdpc\" (UniqueName: \"kubernetes.io/projected/c6c66668-1556-427e-be1d-1d36a93e1369-kube-api-access-rgdpc\") pod \"watcher-kuttl-db-sync-6s9wh\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:48 crc kubenswrapper[4807]: I1205 12:31:48.092912 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:48 crc kubenswrapper[4807]: W1205 12:31:48.610416 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6c66668_1556_427e_be1d_1d36a93e1369.slice/crio-1df8266d1e4633a353f930da2ab8228a2aa27cf6e7fea50350e4cf22d830c7aa WatchSource:0}: Error finding container 1df8266d1e4633a353f930da2ab8228a2aa27cf6e7fea50350e4cf22d830c7aa: Status 404 returned error can't find the container with id 1df8266d1e4633a353f930da2ab8228a2aa27cf6e7fea50350e4cf22d830c7aa Dec 05 12:31:48 crc kubenswrapper[4807]: I1205 12:31:48.610836 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh"] Dec 05 12:31:49 crc kubenswrapper[4807]: I1205 12:31:49.176297 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerStarted","Data":"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557"} Dec 05 12:31:49 crc kubenswrapper[4807]: I1205 12:31:49.177002 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:31:49 crc kubenswrapper[4807]: I1205 12:31:49.178897 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" event={"ID":"c6c66668-1556-427e-be1d-1d36a93e1369","Type":"ContainerStarted","Data":"56d4f3dc25a1baabfa69697662d7947396c5b37052889719cb8740cd7db4b654"} Dec 05 12:31:49 crc kubenswrapper[4807]: I1205 12:31:49.178922 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" event={"ID":"c6c66668-1556-427e-be1d-1d36a93e1369","Type":"ContainerStarted","Data":"1df8266d1e4633a353f930da2ab8228a2aa27cf6e7fea50350e4cf22d830c7aa"} Dec 05 12:31:49 crc kubenswrapper[4807]: I1205 12:31:49.209339 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.140574518 podStartE2EDuration="6.209319944s" podCreationTimestamp="2025-12-05 12:31:43 +0000 UTC" firstStartedPulling="2025-12-05 12:31:44.017257322 +0000 UTC m=+1533.511120591" lastFinishedPulling="2025-12-05 12:31:48.086002748 +0000 UTC m=+1537.579866017" observedRunningTime="2025-12-05 12:31:49.200662713 +0000 UTC m=+1538.694526002" watchObservedRunningTime="2025-12-05 12:31:49.209319944 +0000 UTC m=+1538.703183213" Dec 05 12:31:49 crc kubenswrapper[4807]: I1205 12:31:49.223729 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" podStartSLOduration=2.223709875 podStartE2EDuration="2.223709875s" podCreationTimestamp="2025-12-05 12:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:49.219898932 +0000 UTC m=+1538.713762201" watchObservedRunningTime="2025-12-05 12:31:49.223709875 +0000 UTC m=+1538.717573144" Dec 05 12:31:52 crc kubenswrapper[4807]: I1205 12:31:52.203462 4807 generic.go:334] "Generic (PLEG): container finished" podID="c6c66668-1556-427e-be1d-1d36a93e1369" containerID="56d4f3dc25a1baabfa69697662d7947396c5b37052889719cb8740cd7db4b654" exitCode=0 Dec 05 12:31:52 crc kubenswrapper[4807]: I1205 12:31:52.203868 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" event={"ID":"c6c66668-1556-427e-be1d-1d36a93e1369","Type":"ContainerDied","Data":"56d4f3dc25a1baabfa69697662d7947396c5b37052889719cb8740cd7db4b654"} Dec 05 12:31:52 crc kubenswrapper[4807]: I1205 12:31:52.466653 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:31:52 crc kubenswrapper[4807]: I1205 12:31:52.466706 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.621517 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.741628 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-config-data\") pod \"c6c66668-1556-427e-be1d-1d36a93e1369\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.741679 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgdpc\" (UniqueName: \"kubernetes.io/projected/c6c66668-1556-427e-be1d-1d36a93e1369-kube-api-access-rgdpc\") pod \"c6c66668-1556-427e-be1d-1d36a93e1369\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.741712 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-db-sync-config-data\") pod \"c6c66668-1556-427e-be1d-1d36a93e1369\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.741819 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-combined-ca-bundle\") pod \"c6c66668-1556-427e-be1d-1d36a93e1369\" (UID: \"c6c66668-1556-427e-be1d-1d36a93e1369\") " Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.758617 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c6c66668-1556-427e-be1d-1d36a93e1369" (UID: "c6c66668-1556-427e-be1d-1d36a93e1369"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.760542 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6c66668-1556-427e-be1d-1d36a93e1369-kube-api-access-rgdpc" (OuterVolumeSpecName: "kube-api-access-rgdpc") pod "c6c66668-1556-427e-be1d-1d36a93e1369" (UID: "c6c66668-1556-427e-be1d-1d36a93e1369"). InnerVolumeSpecName "kube-api-access-rgdpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.763609 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6c66668-1556-427e-be1d-1d36a93e1369" (UID: "c6c66668-1556-427e-be1d-1d36a93e1369"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.794863 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-config-data" (OuterVolumeSpecName: "config-data") pod "c6c66668-1556-427e-be1d-1d36a93e1369" (UID: "c6c66668-1556-427e-be1d-1d36a93e1369"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.843743 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgdpc\" (UniqueName: \"kubernetes.io/projected/c6c66668-1556-427e-be1d-1d36a93e1369-kube-api-access-rgdpc\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.843774 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.843783 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:53 crc kubenswrapper[4807]: I1205 12:31:53.843792 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6c66668-1556-427e-be1d-1d36a93e1369-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.220997 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" event={"ID":"c6c66668-1556-427e-be1d-1d36a93e1369","Type":"ContainerDied","Data":"1df8266d1e4633a353f930da2ab8228a2aa27cf6e7fea50350e4cf22d830c7aa"} Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.221039 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1df8266d1e4633a353f930da2ab8228a2aa27cf6e7fea50350e4cf22d830c7aa" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.221046 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.563551 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:54 crc kubenswrapper[4807]: E1205 12:31:54.563879 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6c66668-1556-427e-be1d-1d36a93e1369" containerName="watcher-kuttl-db-sync" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.563902 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6c66668-1556-427e-be1d-1d36a93e1369" containerName="watcher-kuttl-db-sync" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.564086 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6c66668-1556-427e-be1d-1d36a93e1369" containerName="watcher-kuttl-db-sync" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.565218 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.574632 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.597119 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.597119 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.597256 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-9zr5r" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.619101 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.629164 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.630212 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: W1205 12:31:54.634099 4807 reflector.go:561] object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data": failed to list *v1.Secret: secrets "watcher-kuttl-applier-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "watcher-kuttl-default": no relationship found between node 'crc' and this object Dec 05 12:31:54 crc kubenswrapper[4807]: E1205 12:31:54.634349 4807 reflector.go:158] "Unhandled Error" err="object-\"watcher-kuttl-default\"/\"watcher-kuttl-applier-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"watcher-kuttl-applier-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"watcher-kuttl-default\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.656684 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.656739 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bgnm\" (UniqueName: \"kubernetes.io/projected/dacbae36-a301-457c-b53f-0b55d0e1c53a-kube-api-access-6bgnm\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.656785 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dacbae36-a301-457c-b53f-0b55d0e1c53a-logs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.656813 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.656868 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.656927 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.656948 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.684641 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.731099 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.741000 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.743924 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.750575 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759327 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vrv7\" (UniqueName: \"kubernetes.io/projected/ca374238-611b-4d98-ac2a-decc44c15643-kube-api-access-7vrv7\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759375 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759403 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759431 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759446 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759473 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759499 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bgnm\" (UniqueName: \"kubernetes.io/projected/dacbae36-a301-457c-b53f-0b55d0e1c53a-kube-api-access-6bgnm\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759544 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dacbae36-a301-457c-b53f-0b55d0e1c53a-logs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759560 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca374238-611b-4d98-ac2a-decc44c15643-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759579 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.759619 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.760583 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dacbae36-a301-457c-b53f-0b55d0e1c53a-logs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.768916 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.770284 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.772447 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.775769 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.781385 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.794200 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bgnm\" (UniqueName: \"kubernetes.io/projected/dacbae36-a301-457c-b53f-0b55d0e1c53a-kube-api-access-6bgnm\") pod \"watcher-kuttl-api-0\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865439 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865514 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca374238-611b-4d98-ac2a-decc44c15643-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865565 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e70517a-8efe-4fd6-a13a-04f149dedf10-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865597 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865681 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vrv7\" (UniqueName: \"kubernetes.io/projected/ca374238-611b-4d98-ac2a-decc44c15643-kube-api-access-7vrv7\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865734 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v86zg\" (UniqueName: \"kubernetes.io/projected/2e70517a-8efe-4fd6-a13a-04f149dedf10-kube-api-access-v86zg\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865758 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865778 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.865814 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.866370 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca374238-611b-4d98-ac2a-decc44c15643-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.869837 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.885900 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.895214 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vrv7\" (UniqueName: \"kubernetes.io/projected/ca374238-611b-4d98-ac2a-decc44c15643-kube-api-access-7vrv7\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.967727 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v86zg\" (UniqueName: \"kubernetes.io/projected/2e70517a-8efe-4fd6-a13a-04f149dedf10-kube-api-access-v86zg\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.967999 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.968149 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.968277 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e70517a-8efe-4fd6-a13a-04f149dedf10-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.968385 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.970000 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e70517a-8efe-4fd6-a13a-04f149dedf10-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.972320 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.972686 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.980658 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:54 crc kubenswrapper[4807]: I1205 12:31:54.999957 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v86zg\" (UniqueName: \"kubernetes.io/projected/2e70517a-8efe-4fd6-a13a-04f149dedf10-kube-api-access-v86zg\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:55 crc kubenswrapper[4807]: I1205 12:31:55.060077 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:31:55 crc kubenswrapper[4807]: W1205 12:31:55.448453 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddacbae36_a301_457c_b53f_0b55d0e1c53a.slice/crio-c3ca561e0460cb2479f2fb1a141dd611d55a336dade3c8a7ea5d640f23704434 WatchSource:0}: Error finding container c3ca561e0460cb2479f2fb1a141dd611d55a336dade3c8a7ea5d640f23704434: Status 404 returned error can't find the container with id c3ca561e0460cb2479f2fb1a141dd611d55a336dade3c8a7ea5d640f23704434 Dec 05 12:31:55 crc kubenswrapper[4807]: I1205 12:31:55.448783 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:31:55 crc kubenswrapper[4807]: I1205 12:31:55.568056 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:31:55 crc kubenswrapper[4807]: W1205 12:31:55.569476 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e70517a_8efe_4fd6_a13a_04f149dedf10.slice/crio-fbbbed0f76bf060344e7c0bf30a2b0747ec7a95c28ce171849249cc90811fd69 WatchSource:0}: Error finding container fbbbed0f76bf060344e7c0bf30a2b0747ec7a95c28ce171849249cc90811fd69: Status 404 returned error can't find the container with id fbbbed0f76bf060344e7c0bf30a2b0747ec7a95c28ce171849249cc90811fd69 Dec 05 12:31:55 crc kubenswrapper[4807]: E1205 12:31:55.866421 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-applier-config-data: failed to sync secret cache: timed out waiting for the condition Dec 05 12:31:55 crc kubenswrapper[4807]: E1205 12:31:55.866869 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data podName:ca374238-611b-4d98-ac2a-decc44c15643 nodeName:}" failed. No retries permitted until 2025-12-05 12:31:56.366845394 +0000 UTC m=+1545.860708663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data") pod "watcher-kuttl-applier-0" (UID: "ca374238-611b-4d98-ac2a-decc44c15643") : failed to sync secret cache: timed out waiting for the condition Dec 05 12:31:55 crc kubenswrapper[4807]: I1205 12:31:55.929496 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.243565 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"2e70517a-8efe-4fd6-a13a-04f149dedf10","Type":"ContainerStarted","Data":"131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d"} Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.243614 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"2e70517a-8efe-4fd6-a13a-04f149dedf10","Type":"ContainerStarted","Data":"fbbbed0f76bf060344e7c0bf30a2b0747ec7a95c28ce171849249cc90811fd69"} Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.247384 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"dacbae36-a301-457c-b53f-0b55d0e1c53a","Type":"ContainerStarted","Data":"0e9af1c33fd02732643ae74fe10b886e68461da942d351774099d9913dd80426"} Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.247429 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"dacbae36-a301-457c-b53f-0b55d0e1c53a","Type":"ContainerStarted","Data":"47f03ec7d56037ab042093a3bffc51e8b418d0644affa1abf63f6b78fe0dabc3"} Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.247442 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"dacbae36-a301-457c-b53f-0b55d0e1c53a","Type":"ContainerStarted","Data":"c3ca561e0460cb2479f2fb1a141dd611d55a336dade3c8a7ea5d640f23704434"} Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.249601 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.267879 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.26785418 podStartE2EDuration="2.26785418s" podCreationTimestamp="2025-12-05 12:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:56.264282182 +0000 UTC m=+1545.758145451" watchObservedRunningTime="2025-12-05 12:31:56.26785418 +0000 UTC m=+1545.761717449" Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.294672 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.2946588820000002 podStartE2EDuration="2.294658882s" podCreationTimestamp="2025-12-05 12:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:56.292251393 +0000 UTC m=+1545.786114672" watchObservedRunningTime="2025-12-05 12:31:56.294658882 +0000 UTC m=+1545.788522151" Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.394952 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.410400 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.450483 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:31:56 crc kubenswrapper[4807]: I1205 12:31:56.920896 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:31:56 crc kubenswrapper[4807]: W1205 12:31:56.921916 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca374238_611b_4d98_ac2a_decc44c15643.slice/crio-c5ba6ee9814dc74e9423178db3e7e159b8657c8249d64000cbe07cdf5e19b7aa WatchSource:0}: Error finding container c5ba6ee9814dc74e9423178db3e7e159b8657c8249d64000cbe07cdf5e19b7aa: Status 404 returned error can't find the container with id c5ba6ee9814dc74e9423178db3e7e159b8657c8249d64000cbe07cdf5e19b7aa Dec 05 12:31:57 crc kubenswrapper[4807]: I1205 12:31:57.257926 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ca374238-611b-4d98-ac2a-decc44c15643","Type":"ContainerStarted","Data":"77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff"} Dec 05 12:31:57 crc kubenswrapper[4807]: I1205 12:31:57.258265 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ca374238-611b-4d98-ac2a-decc44c15643","Type":"ContainerStarted","Data":"c5ba6ee9814dc74e9423178db3e7e159b8657c8249d64000cbe07cdf5e19b7aa"} Dec 05 12:31:57 crc kubenswrapper[4807]: I1205 12:31:57.284868 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=3.284846979 podStartE2EDuration="3.284846979s" podCreationTimestamp="2025-12-05 12:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:31:57.277138422 +0000 UTC m=+1546.771001701" watchObservedRunningTime="2025-12-05 12:31:57.284846979 +0000 UTC m=+1546.778710248" Dec 05 12:31:58 crc kubenswrapper[4807]: I1205 12:31:58.267618 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:31:58 crc kubenswrapper[4807]: I1205 12:31:58.925476 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:31:59 crc kubenswrapper[4807]: I1205 12:31:59.887866 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:01 crc kubenswrapper[4807]: I1205 12:32:01.450970 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:04 crc kubenswrapper[4807]: I1205 12:32:04.887240 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:04 crc kubenswrapper[4807]: I1205 12:32:04.895977 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:05 crc kubenswrapper[4807]: I1205 12:32:05.061398 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:05 crc kubenswrapper[4807]: I1205 12:32:05.104822 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:05 crc kubenswrapper[4807]: I1205 12:32:05.321682 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:05 crc kubenswrapper[4807]: I1205 12:32:05.330365 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:05 crc kubenswrapper[4807]: I1205 12:32:05.350238 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:06 crc kubenswrapper[4807]: I1205 12:32:06.451598 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:06 crc kubenswrapper[4807]: I1205 12:32:06.475995 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:07 crc kubenswrapper[4807]: I1205 12:32:07.366211 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:07 crc kubenswrapper[4807]: I1205 12:32:07.457592 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:32:07 crc kubenswrapper[4807]: I1205 12:32:07.458037 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-central-agent" containerID="cri-o://a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" gracePeriod=30 Dec 05 12:32:07 crc kubenswrapper[4807]: I1205 12:32:07.458097 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="sg-core" containerID="cri-o://66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" gracePeriod=30 Dec 05 12:32:07 crc kubenswrapper[4807]: I1205 12:32:07.458134 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-notification-agent" containerID="cri-o://6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" gracePeriod=30 Dec 05 12:32:07 crc kubenswrapper[4807]: I1205 12:32:07.458082 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="proxy-httpd" containerID="cri-o://2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" gracePeriod=30 Dec 05 12:32:07 crc kubenswrapper[4807]: I1205 12:32:07.467365 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.154:3000/\": EOF" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.320948 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.350422 4807 generic.go:334] "Generic (PLEG): container finished" podID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerID="2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" exitCode=0 Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.350454 4807 generic.go:334] "Generic (PLEG): container finished" podID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerID="66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" exitCode=2 Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.350463 4807 generic.go:334] "Generic (PLEG): container finished" podID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerID="6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" exitCode=0 Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.350470 4807 generic.go:334] "Generic (PLEG): container finished" podID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerID="a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" exitCode=0 Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.351164 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.351742 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerDied","Data":"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557"} Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.351778 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerDied","Data":"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9"} Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.351793 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerDied","Data":"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5"} Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.351806 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerDied","Data":"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac"} Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.351817 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"dd39b4a0-3725-4ae9-bad5-8cf81a720a33","Type":"ContainerDied","Data":"d72e7291d30b9f78288122d5c660d088052512c1c1bb67ed827b91412c84ef9b"} Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.351835 4807 scope.go:117] "RemoveContainer" containerID="2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.389993 4807 scope.go:117] "RemoveContainer" containerID="66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.406593 4807 scope.go:117] "RemoveContainer" containerID="6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.433116 4807 scope.go:117] "RemoveContainer" containerID="a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.455006 4807 scope.go:117] "RemoveContainer" containerID="2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.455472 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": container with ID starting with 2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557 not found: ID does not exist" containerID="2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.455551 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557"} err="failed to get container status \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": rpc error: code = NotFound desc = could not find container \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": container with ID starting with 2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.455584 4807 scope.go:117] "RemoveContainer" containerID="66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.456127 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": container with ID starting with 66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9 not found: ID does not exist" containerID="66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.456163 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9"} err="failed to get container status \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": rpc error: code = NotFound desc = could not find container \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": container with ID starting with 66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.456186 4807 scope.go:117] "RemoveContainer" containerID="6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.456456 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": container with ID starting with 6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5 not found: ID does not exist" containerID="6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.456502 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5"} err="failed to get container status \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": rpc error: code = NotFound desc = could not find container \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": container with ID starting with 6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.456553 4807 scope.go:117] "RemoveContainer" containerID="a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.456866 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": container with ID starting with a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac not found: ID does not exist" containerID="a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.456901 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac"} err="failed to get container status \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": rpc error: code = NotFound desc = could not find container \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": container with ID starting with a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.456920 4807 scope.go:117] "RemoveContainer" containerID="2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457153 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557"} err="failed to get container status \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": rpc error: code = NotFound desc = could not find container \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": container with ID starting with 2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457181 4807 scope.go:117] "RemoveContainer" containerID="66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457379 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9"} err="failed to get container status \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": rpc error: code = NotFound desc = could not find container \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": container with ID starting with 66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457403 4807 scope.go:117] "RemoveContainer" containerID="6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457671 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5"} err="failed to get container status \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": rpc error: code = NotFound desc = could not find container \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": container with ID starting with 6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457704 4807 scope.go:117] "RemoveContainer" containerID="a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457973 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac"} err="failed to get container status \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": rpc error: code = NotFound desc = could not find container \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": container with ID starting with a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.457993 4807 scope.go:117] "RemoveContainer" containerID="2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.458235 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557"} err="failed to get container status \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": rpc error: code = NotFound desc = could not find container \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": container with ID starting with 2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.458253 4807 scope.go:117] "RemoveContainer" containerID="66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.458573 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9"} err="failed to get container status \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": rpc error: code = NotFound desc = could not find container \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": container with ID starting with 66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.458601 4807 scope.go:117] "RemoveContainer" containerID="6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.458863 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5"} err="failed to get container status \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": rpc error: code = NotFound desc = could not find container \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": container with ID starting with 6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.458882 4807 scope.go:117] "RemoveContainer" containerID="a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459145 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac"} err="failed to get container status \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": rpc error: code = NotFound desc = could not find container \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": container with ID starting with a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459166 4807 scope.go:117] "RemoveContainer" containerID="2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459382 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557"} err="failed to get container status \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": rpc error: code = NotFound desc = could not find container \"2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557\": container with ID starting with 2c184ed1ba0b65d64339ec19f827368bd477856d378cabf6e77b6079f3d92557 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459408 4807 scope.go:117] "RemoveContainer" containerID="66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459636 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9"} err="failed to get container status \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": rpc error: code = NotFound desc = could not find container \"66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9\": container with ID starting with 66e68e86def0c4ae37cca413716edde4a8b12e21bfd0467226fabdf33ba29ad9 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459658 4807 scope.go:117] "RemoveContainer" containerID="6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459935 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5"} err="failed to get container status \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": rpc error: code = NotFound desc = could not find container \"6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5\": container with ID starting with 6477551cafe164bd02ac15afcdf21b3be25f434339159422aed75cac0783d9f5 not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.459959 4807 scope.go:117] "RemoveContainer" containerID="a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.460224 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac"} err="failed to get container status \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": rpc error: code = NotFound desc = could not find container \"a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac\": container with ID starting with a64c54047ebf492f3cbdcc2ddf3dab474742afba6a069a6916c650996ccf8bac not found: ID does not exist" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489068 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-combined-ca-bundle\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489394 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-scripts\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489426 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-config-data\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489456 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-log-httpd\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489485 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-sg-core-conf-yaml\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489567 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qggsm\" (UniqueName: \"kubernetes.io/projected/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-kube-api-access-qggsm\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489592 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-ceilometer-tls-certs\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.489643 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-run-httpd\") pod \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\" (UID: \"dd39b4a0-3725-4ae9-bad5-8cf81a720a33\") " Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.490744 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.491065 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.494254 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-scripts" (OuterVolumeSpecName: "scripts") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.495511 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-kube-api-access-qggsm" (OuterVolumeSpecName: "kube-api-access-qggsm") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "kube-api-access-qggsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.517700 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.532280 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.583259 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.586147 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-config-data" (OuterVolumeSpecName: "config-data") pod "dd39b4a0-3725-4ae9-bad5-8cf81a720a33" (UID: "dd39b4a0-3725-4ae9-bad5-8cf81a720a33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592098 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592141 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592154 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592165 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592176 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592188 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qggsm\" (UniqueName: \"kubernetes.io/projected/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-kube-api-access-qggsm\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592203 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.592216 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd39b4a0-3725-4ae9-bad5-8cf81a720a33-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.684438 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.704932 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711154 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.711580 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="proxy-httpd" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711602 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="proxy-httpd" Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.711619 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-notification-agent" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711627 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-notification-agent" Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.711649 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-central-agent" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711656 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-central-agent" Dec 05 12:32:08 crc kubenswrapper[4807]: E1205 12:32:08.711667 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="sg-core" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711672 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="sg-core" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711816 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-notification-agent" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711832 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="sg-core" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711843 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="proxy-httpd" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.711855 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" containerName="ceilometer-central-agent" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.713616 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.716334 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.717139 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.719104 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.732277 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.896211 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-log-httpd\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.896478 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-config-data\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.896639 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.896831 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-scripts\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.896944 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.897046 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.897158 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-run-httpd\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.897286 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbv5g\" (UniqueName: \"kubernetes.io/projected/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-kube-api-access-rbv5g\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999355 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-scripts\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999405 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999427 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999448 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-run-httpd\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999468 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbv5g\" (UniqueName: \"kubernetes.io/projected/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-kube-api-access-rbv5g\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999498 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-log-httpd\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999545 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-config-data\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:08 crc kubenswrapper[4807]: I1205 12:32:08.999563 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.000398 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-run-httpd\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.000442 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-log-httpd\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.004841 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.018293 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.018684 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.019261 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-config-data\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.028389 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbv5g\" (UniqueName: \"kubernetes.io/projected/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-kube-api-access-rbv5g\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.028436 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-scripts\") pod \"ceilometer-0\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.245287 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd39b4a0-3725-4ae9-bad5-8cf81a720a33" path="/var/lib/kubelet/pods/dd39b4a0-3725-4ae9-bad5-8cf81a720a33/volumes" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.329774 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:09 crc kubenswrapper[4807]: I1205 12:32:09.807721 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:32:09 crc kubenswrapper[4807]: W1205 12:32:09.817771 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb7f50a8_b2a9_4a77_92b5_a67a1346c140.slice/crio-55ac9b9c6f5bac726ba0ca2e7f8353c8c25127a378e5eaafbb3c749ced9c9714 WatchSource:0}: Error finding container 55ac9b9c6f5bac726ba0ca2e7f8353c8c25127a378e5eaafbb3c749ced9c9714: Status 404 returned error can't find the container with id 55ac9b9c6f5bac726ba0ca2e7f8353c8c25127a378e5eaafbb3c749ced9c9714 Dec 05 12:32:10 crc kubenswrapper[4807]: I1205 12:32:10.372737 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerStarted","Data":"55ac9b9c6f5bac726ba0ca2e7f8353c8c25127a378e5eaafbb3c749ced9c9714"} Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.384108 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerStarted","Data":"78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65"} Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.384656 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerStarted","Data":"a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a"} Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.667892 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2r8jx"] Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.669485 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.688289 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2r8jx"] Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.747675 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-catalog-content\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.747722 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhjbv\" (UniqueName: \"kubernetes.io/projected/a05d4342-a67e-4cbb-8661-7770e211a848-kube-api-access-jhjbv\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.747759 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-utilities\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.849676 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-catalog-content\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.849716 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhjbv\" (UniqueName: \"kubernetes.io/projected/a05d4342-a67e-4cbb-8661-7770e211a848-kube-api-access-jhjbv\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.849754 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-utilities\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.850244 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-catalog-content\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.850273 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-utilities\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.877009 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhjbv\" (UniqueName: \"kubernetes.io/projected/a05d4342-a67e-4cbb-8661-7770e211a848-kube-api-access-jhjbv\") pod \"community-operators-2r8jx\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:11 crc kubenswrapper[4807]: I1205 12:32:11.989548 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:12 crc kubenswrapper[4807]: I1205 12:32:12.393867 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerStarted","Data":"69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495"} Dec 05 12:32:12 crc kubenswrapper[4807]: I1205 12:32:12.485254 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2r8jx"] Dec 05 12:32:13 crc kubenswrapper[4807]: I1205 12:32:13.413607 4807 generic.go:334] "Generic (PLEG): container finished" podID="a05d4342-a67e-4cbb-8661-7770e211a848" containerID="9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0" exitCode=0 Dec 05 12:32:13 crc kubenswrapper[4807]: I1205 12:32:13.414188 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2r8jx" event={"ID":"a05d4342-a67e-4cbb-8661-7770e211a848","Type":"ContainerDied","Data":"9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0"} Dec 05 12:32:13 crc kubenswrapper[4807]: I1205 12:32:13.414216 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2r8jx" event={"ID":"a05d4342-a67e-4cbb-8661-7770e211a848","Type":"ContainerStarted","Data":"8d3b4e0706ccc90acfb65c2d4f23b20ed9a7f880d7b40ee62c175394402b48e9"} Dec 05 12:32:14 crc kubenswrapper[4807]: I1205 12:32:14.425873 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerStarted","Data":"f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42"} Dec 05 12:32:14 crc kubenswrapper[4807]: I1205 12:32:14.426382 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:14 crc kubenswrapper[4807]: I1205 12:32:14.430650 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2r8jx" event={"ID":"a05d4342-a67e-4cbb-8661-7770e211a848","Type":"ContainerStarted","Data":"109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335"} Dec 05 12:32:14 crc kubenswrapper[4807]: I1205 12:32:14.455762 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.937262157 podStartE2EDuration="6.455743667s" podCreationTimestamp="2025-12-05 12:32:08 +0000 UTC" firstStartedPulling="2025-12-05 12:32:09.820420958 +0000 UTC m=+1559.314284227" lastFinishedPulling="2025-12-05 12:32:13.338902468 +0000 UTC m=+1562.832765737" observedRunningTime="2025-12-05 12:32:14.449817522 +0000 UTC m=+1563.943680801" watchObservedRunningTime="2025-12-05 12:32:14.455743667 +0000 UTC m=+1563.949606936" Dec 05 12:32:15 crc kubenswrapper[4807]: I1205 12:32:15.439863 4807 generic.go:334] "Generic (PLEG): container finished" podID="a05d4342-a67e-4cbb-8661-7770e211a848" containerID="109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335" exitCode=0 Dec 05 12:32:15 crc kubenswrapper[4807]: I1205 12:32:15.439909 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2r8jx" event={"ID":"a05d4342-a67e-4cbb-8661-7770e211a848","Type":"ContainerDied","Data":"109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335"} Dec 05 12:32:17 crc kubenswrapper[4807]: I1205 12:32:17.461048 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2r8jx" event={"ID":"a05d4342-a67e-4cbb-8661-7770e211a848","Type":"ContainerStarted","Data":"2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5"} Dec 05 12:32:21 crc kubenswrapper[4807]: I1205 12:32:21.990818 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:21 crc kubenswrapper[4807]: I1205 12:32:21.992173 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:22 crc kubenswrapper[4807]: I1205 12:32:22.058898 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:22 crc kubenswrapper[4807]: I1205 12:32:22.076459 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2r8jx" podStartSLOduration=7.959236409 podStartE2EDuration="11.076444758s" podCreationTimestamp="2025-12-05 12:32:11 +0000 UTC" firstStartedPulling="2025-12-05 12:32:13.41585757 +0000 UTC m=+1562.909720839" lastFinishedPulling="2025-12-05 12:32:16.533065919 +0000 UTC m=+1566.026929188" observedRunningTime="2025-12-05 12:32:17.492022257 +0000 UTC m=+1566.985885526" watchObservedRunningTime="2025-12-05 12:32:22.076444758 +0000 UTC m=+1571.570308017" Dec 05 12:32:22 crc kubenswrapper[4807]: I1205 12:32:22.465889 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:32:22 crc kubenswrapper[4807]: I1205 12:32:22.465957 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:32:22 crc kubenswrapper[4807]: I1205 12:32:22.555334 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.831484 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.831746 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="2e70517a-8efe-4fd6-a13a-04f149dedf10" containerName="watcher-decision-engine" containerID="cri-o://131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" gracePeriod=30 Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.843367 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.846258 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/memcached-0" podUID="c9b15b84-7f89-472d-8560-86d63c73b30b" containerName="memcached" containerID="cri-o://a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184" gracePeriod=30 Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.933008 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.933215 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="ca374238-611b-4d98-ac2a-decc44c15643" containerName="watcher-applier" containerID="cri-o://77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff" gracePeriod=30 Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.987568 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.988085 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-kuttl-api-log" containerID="cri-o://47f03ec7d56037ab042093a3bffc51e8b418d0644affa1abf63f6b78fe0dabc3" gracePeriod=30 Dec 05 12:32:23 crc kubenswrapper[4807]: I1205 12:32:23.988185 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-api" containerID="cri-o://0e9af1c33fd02732643ae74fe10b886e68461da942d351774099d9913dd80426" gracePeriod=30 Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.029840 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-hznkk"] Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.037698 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-hznkk"] Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.140316 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-fk576"] Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.141947 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.146978 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"osp-secret" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.147293 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-memcached-mtls" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.166467 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-fk576"] Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.249702 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-config-data\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.249967 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-combined-ca-bundle\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.250087 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-cert-memcached-mtls\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.250195 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-scripts\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.250315 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6t8h\" (UniqueName: \"kubernetes.io/projected/15cb5d7c-92f9-4a3c-b279-a1b36844c133-kube-api-access-w6t8h\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.250474 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-credential-keys\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.250615 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-fernet-keys\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.352641 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-fernet-keys\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.354116 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-config-data\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.354319 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-combined-ca-bundle\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.354462 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-cert-memcached-mtls\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.354624 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-scripts\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.354764 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6t8h\" (UniqueName: \"kubernetes.io/projected/15cb5d7c-92f9-4a3c-b279-a1b36844c133-kube-api-access-w6t8h\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.355042 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-credential-keys\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.359457 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-credential-keys\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.359928 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-fernet-keys\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.359946 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-combined-ca-bundle\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.360325 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-config-data\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.360594 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-cert-memcached-mtls\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.361587 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-scripts\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.380280 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6t8h\" (UniqueName: \"kubernetes.io/projected/15cb5d7c-92f9-4a3c-b279-a1b36844c133-kube-api-access-w6t8h\") pod \"keystone-bootstrap-fk576\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.469049 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.530268 4807 generic.go:334] "Generic (PLEG): container finished" podID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerID="0e9af1c33fd02732643ae74fe10b886e68461da942d351774099d9913dd80426" exitCode=0 Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.530298 4807 generic.go:334] "Generic (PLEG): container finished" podID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerID="47f03ec7d56037ab042093a3bffc51e8b418d0644affa1abf63f6b78fe0dabc3" exitCode=143 Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.531096 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"dacbae36-a301-457c-b53f-0b55d0e1c53a","Type":"ContainerDied","Data":"0e9af1c33fd02732643ae74fe10b886e68461da942d351774099d9913dd80426"} Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.531126 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"dacbae36-a301-457c-b53f-0b55d0e1c53a","Type":"ContainerDied","Data":"47f03ec7d56037ab042093a3bffc51e8b418d0644affa1abf63f6b78fe0dabc3"} Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.901009 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:24 crc kubenswrapper[4807]: I1205 12:32:24.908948 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-fk576"] Dec 05 12:32:24 crc kubenswrapper[4807]: W1205 12:32:24.912376 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15cb5d7c_92f9_4a3c_b279_a1b36844c133.slice/crio-948434443c160d38d09f5bc5ba1d0182f786de9a564379f30ca1486500a67ee9 WatchSource:0}: Error finding container 948434443c160d38d09f5bc5ba1d0182f786de9a564379f30ca1486500a67ee9: Status 404 returned error can't find the container with id 948434443c160d38d09f5bc5ba1d0182f786de9a564379f30ca1486500a67ee9 Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.065858 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dacbae36-a301-457c-b53f-0b55d0e1c53a-logs\") pod \"dacbae36-a301-457c-b53f-0b55d0e1c53a\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066311 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-combined-ca-bundle\") pod \"dacbae36-a301-457c-b53f-0b55d0e1c53a\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066355 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bgnm\" (UniqueName: \"kubernetes.io/projected/dacbae36-a301-457c-b53f-0b55d0e1c53a-kube-api-access-6bgnm\") pod \"dacbae36-a301-457c-b53f-0b55d0e1c53a\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066388 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-config-data\") pod \"dacbae36-a301-457c-b53f-0b55d0e1c53a\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066439 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-custom-prometheus-ca\") pod \"dacbae36-a301-457c-b53f-0b55d0e1c53a\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066556 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-internal-tls-certs\") pod \"dacbae36-a301-457c-b53f-0b55d0e1c53a\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066582 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-public-tls-certs\") pod \"dacbae36-a301-457c-b53f-0b55d0e1c53a\" (UID: \"dacbae36-a301-457c-b53f-0b55d0e1c53a\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066571 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dacbae36-a301-457c-b53f-0b55d0e1c53a-logs" (OuterVolumeSpecName: "logs") pod "dacbae36-a301-457c-b53f-0b55d0e1c53a" (UID: "dacbae36-a301-457c-b53f-0b55d0e1c53a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.066912 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dacbae36-a301-457c-b53f-0b55d0e1c53a-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: E1205 12:32:25.066027 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Dec 05 12:32:25 crc kubenswrapper[4807]: E1205 12:32:25.069010 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Dec 05 12:32:25 crc kubenswrapper[4807]: E1205 12:32:25.070249 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Dec 05 12:32:25 crc kubenswrapper[4807]: E1205 12:32:25.070321 4807 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="2e70517a-8efe-4fd6-a13a-04f149dedf10" containerName="watcher-decision-engine" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.071543 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dacbae36-a301-457c-b53f-0b55d0e1c53a-kube-api-access-6bgnm" (OuterVolumeSpecName: "kube-api-access-6bgnm") pod "dacbae36-a301-457c-b53f-0b55d0e1c53a" (UID: "dacbae36-a301-457c-b53f-0b55d0e1c53a"). InnerVolumeSpecName "kube-api-access-6bgnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.103014 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dacbae36-a301-457c-b53f-0b55d0e1c53a" (UID: "dacbae36-a301-457c-b53f-0b55d0e1c53a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.104632 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "dacbae36-a301-457c-b53f-0b55d0e1c53a" (UID: "dacbae36-a301-457c-b53f-0b55d0e1c53a"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.119951 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dacbae36-a301-457c-b53f-0b55d0e1c53a" (UID: "dacbae36-a301-457c-b53f-0b55d0e1c53a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.137920 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dacbae36-a301-457c-b53f-0b55d0e1c53a" (UID: "dacbae36-a301-457c-b53f-0b55d0e1c53a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.145825 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-config-data" (OuterVolumeSpecName: "config-data") pod "dacbae36-a301-457c-b53f-0b55d0e1c53a" (UID: "dacbae36-a301-457c-b53f-0b55d0e1c53a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.175003 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.175043 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bgnm\" (UniqueName: \"kubernetes.io/projected/dacbae36-a301-457c-b53f-0b55d0e1c53a-kube-api-access-6bgnm\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.175057 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.175068 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.175083 4807 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.175093 4807 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dacbae36-a301-457c-b53f-0b55d0e1c53a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.249727 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b49ba77-5887-4fb4-a553-8b911154f241" path="/var/lib/kubelet/pods/1b49ba77-5887-4fb4-a553-8b911154f241/volumes" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.539653 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-fk576" event={"ID":"15cb5d7c-92f9-4a3c-b279-a1b36844c133","Type":"ContainerStarted","Data":"fdc8938a67fc44474596e2b1498b97f89d2bd91f7de957188d5834956d170b3c"} Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.540316 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-fk576" event={"ID":"15cb5d7c-92f9-4a3c-b279-a1b36844c133","Type":"ContainerStarted","Data":"948434443c160d38d09f5bc5ba1d0182f786de9a564379f30ca1486500a67ee9"} Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.541321 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"dacbae36-a301-457c-b53f-0b55d0e1c53a","Type":"ContainerDied","Data":"c3ca561e0460cb2479f2fb1a141dd611d55a336dade3c8a7ea5d640f23704434"} Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.541375 4807 scope.go:117] "RemoveContainer" containerID="0e9af1c33fd02732643ae74fe10b886e68461da942d351774099d9913dd80426" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.541399 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.566576 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-bootstrap-fk576" podStartSLOduration=1.5665588179999999 podStartE2EDuration="1.566558818s" podCreationTimestamp="2025-12-05 12:32:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:32:25.558119412 +0000 UTC m=+1575.051982681" watchObservedRunningTime="2025-12-05 12:32:25.566558818 +0000 UTC m=+1575.060422087" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.636998 4807 scope.go:117] "RemoveContainer" containerID="47f03ec7d56037ab042093a3bffc51e8b418d0644affa1abf63f6b78fe0dabc3" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.651568 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.658206 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.696092 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:25 crc kubenswrapper[4807]: E1205 12:32:25.696595 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-kuttl-api-log" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.696618 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-kuttl-api-log" Dec 05 12:32:25 crc kubenswrapper[4807]: E1205 12:32:25.696650 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-api" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.696658 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-api" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.696853 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-kuttl-api-log" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.696879 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-api" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.697986 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.700779 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.701034 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.701277 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.704877 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.788589 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.788687 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50129dd-5851-4f74-9d51-7d0748209cf4-logs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.788715 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.789057 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.789195 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.789285 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.789328 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.789423 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgxgm\" (UniqueName: \"kubernetes.io/projected/a50129dd-5851-4f74-9d51-7d0748209cf4-kube-api-access-vgxgm\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.871990 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2r8jx"] Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.872268 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2r8jx" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="registry-server" containerID="cri-o://2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5" gracePeriod=2 Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890404 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890469 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgxgm\" (UniqueName: \"kubernetes.io/projected/a50129dd-5851-4f74-9d51-7d0748209cf4-kube-api-access-vgxgm\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890563 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890635 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50129dd-5851-4f74-9d51-7d0748209cf4-logs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890664 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890750 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890789 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.890816 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.891649 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50129dd-5851-4f74-9d51-7d0748209cf4-logs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.897402 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.897410 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.898818 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.899029 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.899087 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.902585 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.902898 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.911667 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgxgm\" (UniqueName: \"kubernetes.io/projected/a50129dd-5851-4f74-9d51-7d0748209cf4-kube-api-access-vgxgm\") pod \"watcher-kuttl-api-0\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.994069 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-kolla-config\") pod \"c9b15b84-7f89-472d-8560-86d63c73b30b\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.994139 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-memcached-tls-certs\") pod \"c9b15b84-7f89-472d-8560-86d63c73b30b\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.994163 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-combined-ca-bundle\") pod \"c9b15b84-7f89-472d-8560-86d63c73b30b\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.994213 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxh7v\" (UniqueName: \"kubernetes.io/projected/c9b15b84-7f89-472d-8560-86d63c73b30b-kube-api-access-cxh7v\") pod \"c9b15b84-7f89-472d-8560-86d63c73b30b\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.994271 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-config-data\") pod \"c9b15b84-7f89-472d-8560-86d63c73b30b\" (UID: \"c9b15b84-7f89-472d-8560-86d63c73b30b\") " Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.994511 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "c9b15b84-7f89-472d-8560-86d63c73b30b" (UID: "c9b15b84-7f89-472d-8560-86d63c73b30b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.994929 4807 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:25 crc kubenswrapper[4807]: I1205 12:32:25.995037 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-config-data" (OuterVolumeSpecName: "config-data") pod "c9b15b84-7f89-472d-8560-86d63c73b30b" (UID: "c9b15b84-7f89-472d-8560-86d63c73b30b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.004278 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b15b84-7f89-472d-8560-86d63c73b30b-kube-api-access-cxh7v" (OuterVolumeSpecName: "kube-api-access-cxh7v") pod "c9b15b84-7f89-472d-8560-86d63c73b30b" (UID: "c9b15b84-7f89-472d-8560-86d63c73b30b"). InnerVolumeSpecName "kube-api-access-cxh7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.031047 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.031728 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9b15b84-7f89-472d-8560-86d63c73b30b" (UID: "c9b15b84-7f89-472d-8560-86d63c73b30b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.067728 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "c9b15b84-7f89-472d-8560-86d63c73b30b" (UID: "c9b15b84-7f89-472d-8560-86d63c73b30b"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.095971 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.096003 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxh7v\" (UniqueName: \"kubernetes.io/projected/c9b15b84-7f89-472d-8560-86d63c73b30b-kube-api-access-cxh7v\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.096016 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9b15b84-7f89-472d-8560-86d63c73b30b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.096025 4807 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9b15b84-7f89-472d-8560-86d63c73b30b-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.448759 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.455652 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.459622 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.462806 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.462879 4807 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="ca374238-611b-4d98-ac2a-decc44c15643" containerName="watcher-applier" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.565152 4807 generic.go:334] "Generic (PLEG): container finished" podID="c9b15b84-7f89-472d-8560-86d63c73b30b" containerID="a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184" exitCode=0 Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.565217 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"c9b15b84-7f89-472d-8560-86d63c73b30b","Type":"ContainerDied","Data":"a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184"} Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.565243 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"c9b15b84-7f89-472d-8560-86d63c73b30b","Type":"ContainerDied","Data":"9da67cc6fdb097f46a6767ff5e8975a4c46a47aa22951090af55740a81226114"} Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.565260 4807 scope.go:117] "RemoveContainer" containerID="a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.565270 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.582787 4807 generic.go:334] "Generic (PLEG): container finished" podID="a05d4342-a67e-4cbb-8661-7770e211a848" containerID="2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5" exitCode=0 Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.582897 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2r8jx" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.582897 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2r8jx" event={"ID":"a05d4342-a67e-4cbb-8661-7770e211a848","Type":"ContainerDied","Data":"2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5"} Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.583064 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2r8jx" event={"ID":"a05d4342-a67e-4cbb-8661-7770e211a848","Type":"ContainerDied","Data":"8d3b4e0706ccc90acfb65c2d4f23b20ed9a7f880d7b40ee62c175394402b48e9"} Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.603723 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhjbv\" (UniqueName: \"kubernetes.io/projected/a05d4342-a67e-4cbb-8661-7770e211a848-kube-api-access-jhjbv\") pod \"a05d4342-a67e-4cbb-8661-7770e211a848\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.604129 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-catalog-content\") pod \"a05d4342-a67e-4cbb-8661-7770e211a848\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.604219 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-utilities\") pod \"a05d4342-a67e-4cbb-8661-7770e211a848\" (UID: \"a05d4342-a67e-4cbb-8661-7770e211a848\") " Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.610371 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-utilities" (OuterVolumeSpecName: "utilities") pod "a05d4342-a67e-4cbb-8661-7770e211a848" (UID: "a05d4342-a67e-4cbb-8661-7770e211a848"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.625739 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a05d4342-a67e-4cbb-8661-7770e211a848-kube-api-access-jhjbv" (OuterVolumeSpecName: "kube-api-access-jhjbv") pod "a05d4342-a67e-4cbb-8661-7770e211a848" (UID: "a05d4342-a67e-4cbb-8661-7770e211a848"). InnerVolumeSpecName "kube-api-access-jhjbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.633717 4807 scope.go:117] "RemoveContainer" containerID="a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.638456 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184\": container with ID starting with a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184 not found: ID does not exist" containerID="a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.638535 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184"} err="failed to get container status \"a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184\": rpc error: code = NotFound desc = could not find container \"a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184\": container with ID starting with a8361caba6fd6c982f4c70c6629228f7328388bb3e54bb2d795753f303b3d184 not found: ID does not exist" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.638569 4807 scope.go:117] "RemoveContainer" containerID="2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.639100 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.657633 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.667268 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.674602 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.675052 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b15b84-7f89-472d-8560-86d63c73b30b" containerName="memcached" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.675071 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b15b84-7f89-472d-8560-86d63c73b30b" containerName="memcached" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.675089 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="extract-content" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.675098 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="extract-content" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.675122 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="registry-server" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.675131 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="registry-server" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.675157 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="extract-utilities" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.675165 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="extract-utilities" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.675385 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" containerName="registry-server" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.675409 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b15b84-7f89-472d-8560-86d63c73b30b" containerName="memcached" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.680012 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.680138 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.686278 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-memcached-svc" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.686806 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"memcached-config-data" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.688684 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"memcached-memcached-dockercfg-5rcwl" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.708217 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c483e884-62c1-43bf-9428-5b707b5a2cad-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.708324 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c483e884-62c1-43bf-9428-5b707b5a2cad-config-data\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.708367 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c483e884-62c1-43bf-9428-5b707b5a2cad-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.708402 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5qfj\" (UniqueName: \"kubernetes.io/projected/c483e884-62c1-43bf-9428-5b707b5a2cad-kube-api-access-h5qfj\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.708768 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c483e884-62c1-43bf-9428-5b707b5a2cad-kolla-config\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.708968 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.708997 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhjbv\" (UniqueName: \"kubernetes.io/projected/a05d4342-a67e-4cbb-8661-7770e211a848-kube-api-access-jhjbv\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.711666 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a05d4342-a67e-4cbb-8661-7770e211a848" (UID: "a05d4342-a67e-4cbb-8661-7770e211a848"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.752451 4807 scope.go:117] "RemoveContainer" containerID="109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.797227 4807 scope.go:117] "RemoveContainer" containerID="9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.813012 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c483e884-62c1-43bf-9428-5b707b5a2cad-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.813082 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c483e884-62c1-43bf-9428-5b707b5a2cad-config-data\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.813117 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c483e884-62c1-43bf-9428-5b707b5a2cad-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.813139 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5qfj\" (UniqueName: \"kubernetes.io/projected/c483e884-62c1-43bf-9428-5b707b5a2cad-kube-api-access-h5qfj\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.813250 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c483e884-62c1-43bf-9428-5b707b5a2cad-kolla-config\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.813322 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a05d4342-a67e-4cbb-8661-7770e211a848-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.814163 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c483e884-62c1-43bf-9428-5b707b5a2cad-kolla-config\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.816633 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c483e884-62c1-43bf-9428-5b707b5a2cad-config-data\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.818776 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c483e884-62c1-43bf-9428-5b707b5a2cad-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.820077 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c483e884-62c1-43bf-9428-5b707b5a2cad-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.831660 4807 scope.go:117] "RemoveContainer" containerID="2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.832549 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5\": container with ID starting with 2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5 not found: ID does not exist" containerID="2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.832601 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5"} err="failed to get container status \"2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5\": rpc error: code = NotFound desc = could not find container \"2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5\": container with ID starting with 2f8c323d1afb530748a0bc89ef16a470c540d4f5c9ac249d6d250333ba1248e5 not found: ID does not exist" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.832636 4807 scope.go:117] "RemoveContainer" containerID="109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.834338 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335\": container with ID starting with 109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335 not found: ID does not exist" containerID="109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.834371 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335"} err="failed to get container status \"109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335\": rpc error: code = NotFound desc = could not find container \"109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335\": container with ID starting with 109fd051880f595b4681eaa19f418b2b54cba4d9dcea56335bf5bb62d7b79335 not found: ID does not exist" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.834391 4807 scope.go:117] "RemoveContainer" containerID="9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.835660 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0\": container with ID starting with 9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0 not found: ID does not exist" containerID="9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.835709 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0"} err="failed to get container status \"9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0\": rpc error: code = NotFound desc = could not find container \"9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0\": container with ID starting with 9fe6822d4d01f4d8a40e6973cc590a3f3f6515cea3844c68006a4fc08d45f6b0 not found: ID does not exist" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.835986 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5qfj\" (UniqueName: \"kubernetes.io/projected/c483e884-62c1-43bf-9428-5b707b5a2cad-kube-api-access-h5qfj\") pod \"memcached-0\" (UID: \"c483e884-62c1-43bf-9428-5b707b5a2cad\") " pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:26 crc kubenswrapper[4807]: E1205 12:32:26.846751 4807 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9b15b84_7f89_472d_8560_86d63c73b30b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9b15b84_7f89_472d_8560_86d63c73b30b.slice/crio-9da67cc6fdb097f46a6767ff5e8975a4c46a47aa22951090af55740a81226114\": RecentStats: unable to find data in memory cache]" Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.958598 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2r8jx"] Dec 05 12:32:26 crc kubenswrapper[4807]: I1205 12:32:26.969725 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2r8jx"] Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.046214 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.258970 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a05d4342-a67e-4cbb-8661-7770e211a848" path="/var/lib/kubelet/pods/a05d4342-a67e-4cbb-8661-7770e211a848/volumes" Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.266918 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9b15b84-7f89-472d-8560-86d63c73b30b" path="/var/lib/kubelet/pods/c9b15b84-7f89-472d-8560-86d63c73b30b/volumes" Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.267914 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" path="/var/lib/kubelet/pods/dacbae36-a301-457c-b53f-0b55d0e1c53a/volumes" Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.572684 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.595565 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"a50129dd-5851-4f74-9d51-7d0748209cf4","Type":"ContainerStarted","Data":"3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e"} Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.595600 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"a50129dd-5851-4f74-9d51-7d0748209cf4","Type":"ContainerStarted","Data":"fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636"} Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.595615 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"a50129dd-5851-4f74-9d51-7d0748209cf4","Type":"ContainerStarted","Data":"a6a18fa40b5da85ce67c5f69c48e8e4381e7a6612e2a8a1139f2440715582456"} Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.595955 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.597954 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"c483e884-62c1-43bf-9428-5b707b5a2cad","Type":"ContainerStarted","Data":"541a4a260d7f4916b859fb0dc9f6f8440714f4c2a0241dc498a8573dd9f30fc4"} Dec 05 12:32:27 crc kubenswrapper[4807]: I1205 12:32:27.625412 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.625392001 podStartE2EDuration="2.625392001s" podCreationTimestamp="2025-12-05 12:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:32:27.618470913 +0000 UTC m=+1577.112334182" watchObservedRunningTime="2025-12-05 12:32:27.625392001 +0000 UTC m=+1577.119255270" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.612587 4807 generic.go:334] "Generic (PLEG): container finished" podID="ca374238-611b-4d98-ac2a-decc44c15643" containerID="77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff" exitCode=0 Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.613928 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ca374238-611b-4d98-ac2a-decc44c15643","Type":"ContainerDied","Data":"77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff"} Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.615226 4807 generic.go:334] "Generic (PLEG): container finished" podID="15cb5d7c-92f9-4a3c-b279-a1b36844c133" containerID="fdc8938a67fc44474596e2b1498b97f89d2bd91f7de957188d5834956d170b3c" exitCode=0 Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.615260 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-fk576" event={"ID":"15cb5d7c-92f9-4a3c-b279-a1b36844c133","Type":"ContainerDied","Data":"fdc8938a67fc44474596e2b1498b97f89d2bd91f7de957188d5834956d170b3c"} Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.619076 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"c483e884-62c1-43bf-9428-5b707b5a2cad","Type":"ContainerStarted","Data":"7c4b79d36f3b2680678ffc768ce8be50fd30737b6cbb23cda17e8b7f06e91b7b"} Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.619258 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.660266 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/memcached-0" podStartSLOduration=2.660242035 podStartE2EDuration="2.660242035s" podCreationTimestamp="2025-12-05 12:32:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:32:28.653158702 +0000 UTC m=+1578.147021981" watchObservedRunningTime="2025-12-05 12:32:28.660242035 +0000 UTC m=+1578.154105304" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.698826 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.749083 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data\") pod \"ca374238-611b-4d98-ac2a-decc44c15643\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.749181 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-combined-ca-bundle\") pod \"ca374238-611b-4d98-ac2a-decc44c15643\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.749235 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vrv7\" (UniqueName: \"kubernetes.io/projected/ca374238-611b-4d98-ac2a-decc44c15643-kube-api-access-7vrv7\") pod \"ca374238-611b-4d98-ac2a-decc44c15643\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.749312 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca374238-611b-4d98-ac2a-decc44c15643-logs\") pod \"ca374238-611b-4d98-ac2a-decc44c15643\" (UID: \"ca374238-611b-4d98-ac2a-decc44c15643\") " Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.749810 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca374238-611b-4d98-ac2a-decc44c15643-logs" (OuterVolumeSpecName: "logs") pod "ca374238-611b-4d98-ac2a-decc44c15643" (UID: "ca374238-611b-4d98-ac2a-decc44c15643"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.759189 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca374238-611b-4d98-ac2a-decc44c15643-kube-api-access-7vrv7" (OuterVolumeSpecName: "kube-api-access-7vrv7") pod "ca374238-611b-4d98-ac2a-decc44c15643" (UID: "ca374238-611b-4d98-ac2a-decc44c15643"). InnerVolumeSpecName "kube-api-access-7vrv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.787889 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca374238-611b-4d98-ac2a-decc44c15643" (UID: "ca374238-611b-4d98-ac2a-decc44c15643"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.801457 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data" (OuterVolumeSpecName: "config-data") pod "ca374238-611b-4d98-ac2a-decc44c15643" (UID: "ca374238-611b-4d98-ac2a-decc44c15643"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.851327 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca374238-611b-4d98-ac2a-decc44c15643-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.851370 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.851384 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca374238-611b-4d98-ac2a-decc44c15643-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:28 crc kubenswrapper[4807]: I1205 12:32:28.851400 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vrv7\" (UniqueName: \"kubernetes.io/projected/ca374238-611b-4d98-ac2a-decc44c15643-kube-api-access-7vrv7\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.276327 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mqpvz"] Dec 05 12:32:29 crc kubenswrapper[4807]: E1205 12:32:29.277368 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca374238-611b-4d98-ac2a-decc44c15643" containerName="watcher-applier" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.277439 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca374238-611b-4d98-ac2a-decc44c15643" containerName="watcher-applier" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.277676 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca374238-611b-4d98-ac2a-decc44c15643" containerName="watcher-applier" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.312990 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqpvz"] Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.313128 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.364791 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kktx5\" (UniqueName: \"kubernetes.io/projected/a33c345f-a76e-46c7-a711-f26320370541-kube-api-access-kktx5\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.365110 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-utilities\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.365234 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-catalog-content\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.466420 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-utilities\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.466497 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-catalog-content\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.466560 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kktx5\" (UniqueName: \"kubernetes.io/projected/a33c345f-a76e-46c7-a711-f26320370541-kube-api-access-kktx5\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.467133 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-utilities\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.467213 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-catalog-content\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.493930 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kktx5\" (UniqueName: \"kubernetes.io/projected/a33c345f-a76e-46c7-a711-f26320370541-kube-api-access-kktx5\") pod \"certified-operators-mqpvz\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.629357 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"ca374238-611b-4d98-ac2a-decc44c15643","Type":"ContainerDied","Data":"c5ba6ee9814dc74e9423178db3e7e159b8657c8249d64000cbe07cdf5e19b7aa"} Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.629423 4807 scope.go:117] "RemoveContainer" containerID="77c97b2a4923b86c46971792273665c46d78c572ccf4a0a45bab1c54a8868aff" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.629386 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.640854 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.668312 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.682554 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.706003 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.707327 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.712604 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.727442 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.771503 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.772691 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.772737 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.772844 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.773001 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnxh7\" (UniqueName: \"kubernetes.io/projected/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-kube-api-access-vnxh7\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.881941 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnxh7\" (UniqueName: \"kubernetes.io/projected/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-kube-api-access-vnxh7\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.882009 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.882030 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.882049 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.882090 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.885624 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.888605 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"https://10.217.0.156:9322/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.889331 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.889392 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="dacbae36-a301-457c-b53f-0b55d0e1c53a" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.156:9322/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.902191 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.903082 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:29 crc kubenswrapper[4807]: I1205 12:32:29.915587 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnxh7\" (UniqueName: \"kubernetes.io/projected/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-kube-api-access-vnxh7\") pod \"watcher-kuttl-applier-0\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.065307 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.091266 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6t8h\" (UniqueName: \"kubernetes.io/projected/15cb5d7c-92f9-4a3c-b279-a1b36844c133-kube-api-access-w6t8h\") pod \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.091686 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-combined-ca-bundle\") pod \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.091735 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-cert-memcached-mtls\") pod \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.091763 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-config-data\") pod \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.091803 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-scripts\") pod \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.091824 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-credential-keys\") pod \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.091914 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-fernet-keys\") pod \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\" (UID: \"15cb5d7c-92f9-4a3c-b279-a1b36844c133\") " Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.106289 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.107357 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-scripts" (OuterVolumeSpecName: "scripts") pod "15cb5d7c-92f9-4a3c-b279-a1b36844c133" (UID: "15cb5d7c-92f9-4a3c-b279-a1b36844c133"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.108237 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "15cb5d7c-92f9-4a3c-b279-a1b36844c133" (UID: "15cb5d7c-92f9-4a3c-b279-a1b36844c133"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.133731 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "15cb5d7c-92f9-4a3c-b279-a1b36844c133" (UID: "15cb5d7c-92f9-4a3c-b279-a1b36844c133"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.140981 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15cb5d7c-92f9-4a3c-b279-a1b36844c133-kube-api-access-w6t8h" (OuterVolumeSpecName: "kube-api-access-w6t8h") pod "15cb5d7c-92f9-4a3c-b279-a1b36844c133" (UID: "15cb5d7c-92f9-4a3c-b279-a1b36844c133"). InnerVolumeSpecName "kube-api-access-w6t8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.144638 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15cb5d7c-92f9-4a3c-b279-a1b36844c133" (UID: "15cb5d7c-92f9-4a3c-b279-a1b36844c133"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.153737 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-config-data" (OuterVolumeSpecName: "config-data") pod "15cb5d7c-92f9-4a3c-b279-a1b36844c133" (UID: "15cb5d7c-92f9-4a3c-b279-a1b36844c133"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.193294 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.193327 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.193338 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.193347 4807 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.193356 4807 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.193366 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6t8h\" (UniqueName: \"kubernetes.io/projected/15cb5d7c-92f9-4a3c-b279-a1b36844c133-kube-api-access-w6t8h\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.235014 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "15cb5d7c-92f9-4a3c-b279-a1b36844c133" (UID: "15cb5d7c-92f9-4a3c-b279-a1b36844c133"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.295995 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/15cb5d7c-92f9-4a3c-b279-a1b36844c133-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.310594 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqpvz"] Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.596375 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.602583 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.637912 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"58a7640d-9cfa-4e9b-9501-5e2b4da9a886","Type":"ContainerStarted","Data":"b6c1b11f5290a9537397ea13e8d5d426ec87efed8952995dfbf3efd6984ffa57"} Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.643318 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqpvz" event={"ID":"a33c345f-a76e-46c7-a711-f26320370541","Type":"ContainerStarted","Data":"b84dfa7dedfe75d648e9c42f579dad31a25d209fc2aaad857c4b16d80cc0e720"} Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.644958 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-fk576" event={"ID":"15cb5d7c-92f9-4a3c-b279-a1b36844c133","Type":"ContainerDied","Data":"948434443c160d38d09f5bc5ba1d0182f786de9a564379f30ca1486500a67ee9"} Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.644991 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="948434443c160d38d09f5bc5ba1d0182f786de9a564379f30ca1486500a67ee9" Dec 05 12:32:30 crc kubenswrapper[4807]: I1205 12:32:30.645136 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-fk576" Dec 05 12:32:31 crc kubenswrapper[4807]: I1205 12:32:31.031939 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:31 crc kubenswrapper[4807]: I1205 12:32:31.247051 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca374238-611b-4d98-ac2a-decc44c15643" path="/var/lib/kubelet/pods/ca374238-611b-4d98-ac2a-decc44c15643/volumes" Dec 05 12:32:31 crc kubenswrapper[4807]: I1205 12:32:31.657990 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"58a7640d-9cfa-4e9b-9501-5e2b4da9a886","Type":"ContainerStarted","Data":"51dde105a0fac8edef020314a4311b0a44cac5609616e2b8f5f288192934fb9e"} Dec 05 12:32:31 crc kubenswrapper[4807]: I1205 12:32:31.665276 4807 generic.go:334] "Generic (PLEG): container finished" podID="a33c345f-a76e-46c7-a711-f26320370541" containerID="ed20aadb1de1fc47ab3f3525e90736202e3eb48cc84125a78cc77afa0fb78ed9" exitCode=0 Dec 05 12:32:31 crc kubenswrapper[4807]: I1205 12:32:31.665324 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqpvz" event={"ID":"a33c345f-a76e-46c7-a711-f26320370541","Type":"ContainerDied","Data":"ed20aadb1de1fc47ab3f3525e90736202e3eb48cc84125a78cc77afa0fb78ed9"} Dec 05 12:32:31 crc kubenswrapper[4807]: I1205 12:32:31.679997 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.6799767230000002 podStartE2EDuration="2.679976723s" podCreationTimestamp="2025-12-05 12:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:32:31.67824094 +0000 UTC m=+1581.172104209" watchObservedRunningTime="2025-12-05 12:32:31.679976723 +0000 UTC m=+1581.173839992" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.050747 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/memcached-0" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.197344 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-5c74bb55d6-5dfz2"] Dec 05 12:32:32 crc kubenswrapper[4807]: E1205 12:32:32.197791 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15cb5d7c-92f9-4a3c-b279-a1b36844c133" containerName="keystone-bootstrap" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.197814 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="15cb5d7c-92f9-4a3c-b279-a1b36844c133" containerName="keystone-bootstrap" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.198003 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="15cb5d7c-92f9-4a3c-b279-a1b36844c133" containerName="keystone-bootstrap" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.198591 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.220139 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-5c74bb55d6-5dfz2"] Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224311 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-combined-ca-bundle\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224360 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-config-data\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224427 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-fernet-keys\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224445 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-internal-tls-certs\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224476 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-public-tls-certs\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224508 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjpnd\" (UniqueName: \"kubernetes.io/projected/34385b79-da95-4009-82a1-6e927d2f6d5b-kube-api-access-bjpnd\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224538 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-cert-memcached-mtls\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224578 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-scripts\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.224593 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-credential-keys\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326101 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-fernet-keys\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326175 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-internal-tls-certs\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326225 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-public-tls-certs\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326271 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjpnd\" (UniqueName: \"kubernetes.io/projected/34385b79-da95-4009-82a1-6e927d2f6d5b-kube-api-access-bjpnd\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326293 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-cert-memcached-mtls\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326379 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-scripts\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326399 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-credential-keys\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326478 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-combined-ca-bundle\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.326515 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-config-data\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.332953 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-cert-memcached-mtls\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.333105 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-config-data\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.335428 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-public-tls-certs\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.337201 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-fernet-keys\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.338470 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-combined-ca-bundle\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.340648 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-credential-keys\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.341348 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-internal-tls-certs\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.344998 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34385b79-da95-4009-82a1-6e927d2f6d5b-scripts\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.358349 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjpnd\" (UniqueName: \"kubernetes.io/projected/34385b79-da95-4009-82a1-6e927d2f6d5b-kube-api-access-bjpnd\") pod \"keystone-5c74bb55d6-5dfz2\" (UID: \"34385b79-da95-4009-82a1-6e927d2f6d5b\") " pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.520608 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:32 crc kubenswrapper[4807]: I1205 12:32:32.690030 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqpvz" event={"ID":"a33c345f-a76e-46c7-a711-f26320370541","Type":"ContainerStarted","Data":"cad5b12c616ee212ec48112043cc75326d236569096916dfec6beac27cbb5511"} Dec 05 12:32:33 crc kubenswrapper[4807]: I1205 12:32:33.093695 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-5c74bb55d6-5dfz2"] Dec 05 12:32:33 crc kubenswrapper[4807]: I1205 12:32:33.700098 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" event={"ID":"34385b79-da95-4009-82a1-6e927d2f6d5b","Type":"ContainerStarted","Data":"ac3fd3623e83b999205d2668ade4e7b4474011698dffde3b10c3ef577f4291fc"} Dec 05 12:32:33 crc kubenswrapper[4807]: I1205 12:32:33.700427 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" event={"ID":"34385b79-da95-4009-82a1-6e927d2f6d5b","Type":"ContainerStarted","Data":"47dfe48dc1be291acd60963b452215954a6f80c34e42bf0ac0fc46c89a6e104a"} Dec 05 12:32:33 crc kubenswrapper[4807]: I1205 12:32:33.701728 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:32:33 crc kubenswrapper[4807]: I1205 12:32:33.704030 4807 generic.go:334] "Generic (PLEG): container finished" podID="a33c345f-a76e-46c7-a711-f26320370541" containerID="cad5b12c616ee212ec48112043cc75326d236569096916dfec6beac27cbb5511" exitCode=0 Dec 05 12:32:33 crc kubenswrapper[4807]: I1205 12:32:33.704073 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqpvz" event={"ID":"a33c345f-a76e-46c7-a711-f26320370541","Type":"ContainerDied","Data":"cad5b12c616ee212ec48112043cc75326d236569096916dfec6beac27cbb5511"} Dec 05 12:32:33 crc kubenswrapper[4807]: I1205 12:32:33.720297 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" podStartSLOduration=1.720281365 podStartE2EDuration="1.720281365s" podCreationTimestamp="2025-12-05 12:32:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:32:33.71886429 +0000 UTC m=+1583.212727559" watchObservedRunningTime="2025-12-05 12:32:33.720281365 +0000 UTC m=+1583.214144634" Dec 05 12:32:34 crc kubenswrapper[4807]: I1205 12:32:34.723535 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqpvz" event={"ID":"a33c345f-a76e-46c7-a711-f26320370541","Type":"ContainerStarted","Data":"b0ccb1c224769b45f9dc56ca530610fa906198275a234fd41c76f77e86a90660"} Dec 05 12:32:34 crc kubenswrapper[4807]: I1205 12:32:34.747588 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mqpvz" podStartSLOduration=3.256331032 podStartE2EDuration="5.747569824s" podCreationTimestamp="2025-12-05 12:32:29 +0000 UTC" firstStartedPulling="2025-12-05 12:32:31.667383106 +0000 UTC m=+1581.161246375" lastFinishedPulling="2025-12-05 12:32:34.158621908 +0000 UTC m=+1583.652485167" observedRunningTime="2025-12-05 12:32:34.744547581 +0000 UTC m=+1584.238410850" watchObservedRunningTime="2025-12-05 12:32:34.747569824 +0000 UTC m=+1584.241433093" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.106741 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.426914 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.480623 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e70517a-8efe-4fd6-a13a-04f149dedf10-logs\") pod \"2e70517a-8efe-4fd6-a13a-04f149dedf10\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.480930 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v86zg\" (UniqueName: \"kubernetes.io/projected/2e70517a-8efe-4fd6-a13a-04f149dedf10-kube-api-access-v86zg\") pod \"2e70517a-8efe-4fd6-a13a-04f149dedf10\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.481142 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-combined-ca-bundle\") pod \"2e70517a-8efe-4fd6-a13a-04f149dedf10\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.481291 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-config-data\") pod \"2e70517a-8efe-4fd6-a13a-04f149dedf10\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.481391 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-custom-prometheus-ca\") pod \"2e70517a-8efe-4fd6-a13a-04f149dedf10\" (UID: \"2e70517a-8efe-4fd6-a13a-04f149dedf10\") " Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.481128 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e70517a-8efe-4fd6-a13a-04f149dedf10-logs" (OuterVolumeSpecName: "logs") pod "2e70517a-8efe-4fd6-a13a-04f149dedf10" (UID: "2e70517a-8efe-4fd6-a13a-04f149dedf10"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.482117 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e70517a-8efe-4fd6-a13a-04f149dedf10-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.493047 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e70517a-8efe-4fd6-a13a-04f149dedf10-kube-api-access-v86zg" (OuterVolumeSpecName: "kube-api-access-v86zg") pod "2e70517a-8efe-4fd6-a13a-04f149dedf10" (UID: "2e70517a-8efe-4fd6-a13a-04f149dedf10"). InnerVolumeSpecName "kube-api-access-v86zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.509745 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "2e70517a-8efe-4fd6-a13a-04f149dedf10" (UID: "2e70517a-8efe-4fd6-a13a-04f149dedf10"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.536647 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e70517a-8efe-4fd6-a13a-04f149dedf10" (UID: "2e70517a-8efe-4fd6-a13a-04f149dedf10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.546239 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-config-data" (OuterVolumeSpecName: "config-data") pod "2e70517a-8efe-4fd6-a13a-04f149dedf10" (UID: "2e70517a-8efe-4fd6-a13a-04f149dedf10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.584080 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v86zg\" (UniqueName: \"kubernetes.io/projected/2e70517a-8efe-4fd6-a13a-04f149dedf10-kube-api-access-v86zg\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.584128 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.584140 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.584151 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2e70517a-8efe-4fd6-a13a-04f149dedf10-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.746413 4807 generic.go:334] "Generic (PLEG): container finished" podID="2e70517a-8efe-4fd6-a13a-04f149dedf10" containerID="131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" exitCode=0 Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.746498 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.746516 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"2e70517a-8efe-4fd6-a13a-04f149dedf10","Type":"ContainerDied","Data":"131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d"} Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.746613 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"2e70517a-8efe-4fd6-a13a-04f149dedf10","Type":"ContainerDied","Data":"fbbbed0f76bf060344e7c0bf30a2b0747ec7a95c28ce171849249cc90811fd69"} Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.746638 4807 scope.go:117] "RemoveContainer" containerID="131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.764258 4807 scope.go:117] "RemoveContainer" containerID="131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" Dec 05 12:32:35 crc kubenswrapper[4807]: E1205 12:32:35.764882 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d\": container with ID starting with 131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d not found: ID does not exist" containerID="131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.764941 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d"} err="failed to get container status \"131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d\": rpc error: code = NotFound desc = could not find container \"131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d\": container with ID starting with 131cace6f665f1db014f57aacfdc053577372e5f5c0e7bcaddb7bc479d02587d not found: ID does not exist" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.775366 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.782375 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.806198 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:32:35 crc kubenswrapper[4807]: E1205 12:32:35.806615 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e70517a-8efe-4fd6-a13a-04f149dedf10" containerName="watcher-decision-engine" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.806637 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e70517a-8efe-4fd6-a13a-04f149dedf10" containerName="watcher-decision-engine" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.806867 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e70517a-8efe-4fd6-a13a-04f149dedf10" containerName="watcher-decision-engine" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.807579 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.811347 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.818682 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.889204 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.889262 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpkw5\" (UniqueName: \"kubernetes.io/projected/8e820190-88c8-4df3-9832-2d9d0bb914cc-kube-api-access-mpkw5\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.889343 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.889371 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.889400 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.889423 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e820190-88c8-4df3-9832-2d9d0bb914cc-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.993848 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.993917 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.993950 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.993979 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e820190-88c8-4df3-9832-2d9d0bb914cc-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.994053 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.994100 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpkw5\" (UniqueName: \"kubernetes.io/projected/8e820190-88c8-4df3-9832-2d9d0bb914cc-kube-api-access-mpkw5\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.995915 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e820190-88c8-4df3-9832-2d9d0bb914cc-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.999394 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:35 crc kubenswrapper[4807]: I1205 12:32:35.999509 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:35.999955 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.002030 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.011452 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpkw5\" (UniqueName: \"kubernetes.io/projected/8e820190-88c8-4df3-9832-2d9d0bb914cc-kube-api-access-mpkw5\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.032001 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.042418 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.135458 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.581678 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:32:36 crc kubenswrapper[4807]: W1205 12:32:36.582775 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e820190_88c8_4df3_9832_2d9d0bb914cc.slice/crio-03ac910a1a94f42c3fcd4d63423d15258385fe87d9fda90087425a8a2802a161 WatchSource:0}: Error finding container 03ac910a1a94f42c3fcd4d63423d15258385fe87d9fda90087425a8a2802a161: Status 404 returned error can't find the container with id 03ac910a1a94f42c3fcd4d63423d15258385fe87d9fda90087425a8a2802a161 Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.756976 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"8e820190-88c8-4df3-9832-2d9d0bb914cc","Type":"ContainerStarted","Data":"03ac910a1a94f42c3fcd4d63423d15258385fe87d9fda90087425a8a2802a161"} Dec 05 12:32:36 crc kubenswrapper[4807]: I1205 12:32:36.768346 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:37 crc kubenswrapper[4807]: I1205 12:32:37.248255 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e70517a-8efe-4fd6-a13a-04f149dedf10" path="/var/lib/kubelet/pods/2e70517a-8efe-4fd6-a13a-04f149dedf10/volumes" Dec 05 12:32:37 crc kubenswrapper[4807]: I1205 12:32:37.765410 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"8e820190-88c8-4df3-9832-2d9d0bb914cc","Type":"ContainerStarted","Data":"a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba"} Dec 05 12:32:37 crc kubenswrapper[4807]: I1205 12:32:37.791788 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.791768618 podStartE2EDuration="2.791768618s" podCreationTimestamp="2025-12-05 12:32:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:32:37.786330865 +0000 UTC m=+1587.280194154" watchObservedRunningTime="2025-12-05 12:32:37.791768618 +0000 UTC m=+1587.285631887" Dec 05 12:32:39 crc kubenswrapper[4807]: I1205 12:32:39.346581 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:32:39 crc kubenswrapper[4807]: I1205 12:32:39.641998 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:39 crc kubenswrapper[4807]: I1205 12:32:39.642045 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:39 crc kubenswrapper[4807]: I1205 12:32:39.691659 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:39 crc kubenswrapper[4807]: I1205 12:32:39.831162 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:40 crc kubenswrapper[4807]: I1205 12:32:40.106720 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:40 crc kubenswrapper[4807]: I1205 12:32:40.141383 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:40 crc kubenswrapper[4807]: I1205 12:32:40.833961 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:32:42 crc kubenswrapper[4807]: I1205 12:32:42.669439 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:42 crc kubenswrapper[4807]: I1205 12:32:42.670025 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-kuttl-api-log" containerID="cri-o://fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636" gracePeriod=30 Dec 05 12:32:42 crc kubenswrapper[4807]: I1205 12:32:42.670110 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-api" containerID="cri-o://3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e" gracePeriod=30 Dec 05 12:32:42 crc kubenswrapper[4807]: I1205 12:32:42.812421 4807 generic.go:334] "Generic (PLEG): container finished" podID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerID="fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636" exitCode=143 Dec 05 12:32:42 crc kubenswrapper[4807]: I1205 12:32:42.812467 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"a50129dd-5851-4f74-9d51-7d0748209cf4","Type":"ContainerDied","Data":"fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636"} Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.548486 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631061 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-public-tls-certs\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631455 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-internal-tls-certs\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631514 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-config-data\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631564 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-custom-prometheus-ca\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631620 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-cert-memcached-mtls\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631644 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-combined-ca-bundle\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631722 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50129dd-5851-4f74-9d51-7d0748209cf4-logs\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.631761 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgxgm\" (UniqueName: \"kubernetes.io/projected/a50129dd-5851-4f74-9d51-7d0748209cf4-kube-api-access-vgxgm\") pod \"a50129dd-5851-4f74-9d51-7d0748209cf4\" (UID: \"a50129dd-5851-4f74-9d51-7d0748209cf4\") " Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.632226 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a50129dd-5851-4f74-9d51-7d0748209cf4-logs" (OuterVolumeSpecName: "logs") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.654290 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a50129dd-5851-4f74-9d51-7d0748209cf4-kube-api-access-vgxgm" (OuterVolumeSpecName: "kube-api-access-vgxgm") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "kube-api-access-vgxgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.665777 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqpvz"] Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.666012 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mqpvz" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="registry-server" containerID="cri-o://b0ccb1c224769b45f9dc56ca530610fa906198275a234fd41c76f77e86a90660" gracePeriod=2 Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.682792 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.695756 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.723437 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.724075 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.734118 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.734153 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.734164 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50129dd-5851-4f74-9d51-7d0748209cf4-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.734175 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgxgm\" (UniqueName: \"kubernetes.io/projected/a50129dd-5851-4f74-9d51-7d0748209cf4-kube-api-access-vgxgm\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.734185 4807 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.734195 4807 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.738176 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-config-data" (OuterVolumeSpecName: "config-data") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.740215 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "a50129dd-5851-4f74-9d51-7d0748209cf4" (UID: "a50129dd-5851-4f74-9d51-7d0748209cf4"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.821567 4807 generic.go:334] "Generic (PLEG): container finished" podID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerID="3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e" exitCode=0 Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.821609 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"a50129dd-5851-4f74-9d51-7d0748209cf4","Type":"ContainerDied","Data":"3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e"} Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.821624 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.821640 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"a50129dd-5851-4f74-9d51-7d0748209cf4","Type":"ContainerDied","Data":"a6a18fa40b5da85ce67c5f69c48e8e4381e7a6612e2a8a1139f2440715582456"} Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.821658 4807 scope.go:117] "RemoveContainer" containerID="3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.835230 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.835259 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a50129dd-5851-4f74-9d51-7d0748209cf4-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.853882 4807 scope.go:117] "RemoveContainer" containerID="fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.870752 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.880748 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.892438 4807 scope.go:117] "RemoveContainer" containerID="3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e" Dec 05 12:32:43 crc kubenswrapper[4807]: E1205 12:32:43.893754 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e\": container with ID starting with 3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e not found: ID does not exist" containerID="3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.893791 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e"} err="failed to get container status \"3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e\": rpc error: code = NotFound desc = could not find container \"3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e\": container with ID starting with 3af3cf2bed143aef6ac381813964ad5389da73b2f6f6cf8242cccb8ad071d62e not found: ID does not exist" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.893811 4807 scope.go:117] "RemoveContainer" containerID="fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636" Dec 05 12:32:43 crc kubenswrapper[4807]: E1205 12:32:43.894078 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636\": container with ID starting with fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636 not found: ID does not exist" containerID="fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.894111 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636"} err="failed to get container status \"fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636\": rpc error: code = NotFound desc = could not find container \"fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636\": container with ID starting with fdc1387ea87923f724dfdc2b7b3d324ed31329cf4e249445832cdc48c824b636 not found: ID does not exist" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.907327 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:43 crc kubenswrapper[4807]: E1205 12:32:43.907760 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-api" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.907781 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-api" Dec 05 12:32:43 crc kubenswrapper[4807]: E1205 12:32:43.907812 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-kuttl-api-log" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.907820 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-kuttl-api-log" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.908010 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-api" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.908037 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" containerName="watcher-kuttl-api-log" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.909147 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.911825 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.915202 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.938258 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.938668 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.938872 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.939042 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.939180 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98619dc2-c4b1-45e5-8c69-a8410aafa29a-logs\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:43 crc kubenswrapper[4807]: I1205 12:32:43.939376 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlzxw\" (UniqueName: \"kubernetes.io/projected/98619dc2-c4b1-45e5-8c69-a8410aafa29a-kube-api-access-nlzxw\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.040776 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.041189 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.041262 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.041286 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98619dc2-c4b1-45e5-8c69-a8410aafa29a-logs\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.041312 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlzxw\" (UniqueName: \"kubernetes.io/projected/98619dc2-c4b1-45e5-8c69-a8410aafa29a-kube-api-access-nlzxw\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.041414 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.042004 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98619dc2-c4b1-45e5-8c69-a8410aafa29a-logs\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.045238 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.045322 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.046450 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.053688 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.065723 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlzxw\" (UniqueName: \"kubernetes.io/projected/98619dc2-c4b1-45e5-8c69-a8410aafa29a-kube-api-access-nlzxw\") pod \"watcher-kuttl-api-0\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.233416 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.826687 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:32:44 crc kubenswrapper[4807]: I1205 12:32:44.838471 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"98619dc2-c4b1-45e5-8c69-a8410aafa29a","Type":"ContainerStarted","Data":"b998469ef6a72ccec6a784a06a2d00a33c842db4286fa9f6ee281525a57ce573"} Dec 05 12:32:45 crc kubenswrapper[4807]: I1205 12:32:45.247807 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a50129dd-5851-4f74-9d51-7d0748209cf4" path="/var/lib/kubelet/pods/a50129dd-5851-4f74-9d51-7d0748209cf4/volumes" Dec 05 12:32:45 crc kubenswrapper[4807]: I1205 12:32:45.854891 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"98619dc2-c4b1-45e5-8c69-a8410aafa29a","Type":"ContainerStarted","Data":"3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92"} Dec 05 12:32:45 crc kubenswrapper[4807]: I1205 12:32:45.855270 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:45 crc kubenswrapper[4807]: I1205 12:32:45.855308 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"98619dc2-c4b1-45e5-8c69-a8410aafa29a","Type":"ContainerStarted","Data":"2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6"} Dec 05 12:32:45 crc kubenswrapper[4807]: I1205 12:32:45.858002 4807 generic.go:334] "Generic (PLEG): container finished" podID="a33c345f-a76e-46c7-a711-f26320370541" containerID="b0ccb1c224769b45f9dc56ca530610fa906198275a234fd41c76f77e86a90660" exitCode=0 Dec 05 12:32:45 crc kubenswrapper[4807]: I1205 12:32:45.858042 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqpvz" event={"ID":"a33c345f-a76e-46c7-a711-f26320370541","Type":"ContainerDied","Data":"b0ccb1c224769b45f9dc56ca530610fa906198275a234fd41c76f77e86a90660"} Dec 05 12:32:45 crc kubenswrapper[4807]: I1205 12:32:45.871910 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.871888544 podStartE2EDuration="2.871888544s" podCreationTimestamp="2025-12-05 12:32:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:32:45.871283659 +0000 UTC m=+1595.365146928" watchObservedRunningTime="2025-12-05 12:32:45.871888544 +0000 UTC m=+1595.365751803" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.135891 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.165075 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.177228 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.282285 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kktx5\" (UniqueName: \"kubernetes.io/projected/a33c345f-a76e-46c7-a711-f26320370541-kube-api-access-kktx5\") pod \"a33c345f-a76e-46c7-a711-f26320370541\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.282473 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-utilities\") pod \"a33c345f-a76e-46c7-a711-f26320370541\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.282538 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-catalog-content\") pod \"a33c345f-a76e-46c7-a711-f26320370541\" (UID: \"a33c345f-a76e-46c7-a711-f26320370541\") " Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.284094 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-utilities" (OuterVolumeSpecName: "utilities") pod "a33c345f-a76e-46c7-a711-f26320370541" (UID: "a33c345f-a76e-46c7-a711-f26320370541"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.301806 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a33c345f-a76e-46c7-a711-f26320370541-kube-api-access-kktx5" (OuterVolumeSpecName: "kube-api-access-kktx5") pod "a33c345f-a76e-46c7-a711-f26320370541" (UID: "a33c345f-a76e-46c7-a711-f26320370541"). InnerVolumeSpecName "kube-api-access-kktx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.331291 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a33c345f-a76e-46c7-a711-f26320370541" (UID: "a33c345f-a76e-46c7-a711-f26320370541"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.384048 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kktx5\" (UniqueName: \"kubernetes.io/projected/a33c345f-a76e-46c7-a711-f26320370541-kube-api-access-kktx5\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.384091 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.384103 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33c345f-a76e-46c7-a711-f26320370541-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.867262 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqpvz" event={"ID":"a33c345f-a76e-46c7-a711-f26320370541","Type":"ContainerDied","Data":"b84dfa7dedfe75d648e9c42f579dad31a25d209fc2aaad857c4b16d80cc0e720"} Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.867318 4807 scope.go:117] "RemoveContainer" containerID="b0ccb1c224769b45f9dc56ca530610fa906198275a234fd41c76f77e86a90660" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.867374 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqpvz" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.868496 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.909016 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.912481 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqpvz"] Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.915515 4807 scope.go:117] "RemoveContainer" containerID="cad5b12c616ee212ec48112043cc75326d236569096916dfec6beac27cbb5511" Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.919472 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mqpvz"] Dec 05 12:32:46 crc kubenswrapper[4807]: I1205 12:32:46.940842 4807 scope.go:117] "RemoveContainer" containerID="ed20aadb1de1fc47ab3f3525e90736202e3eb48cc84125a78cc77afa0fb78ed9" Dec 05 12:32:47 crc kubenswrapper[4807]: I1205 12:32:47.249087 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a33c345f-a76e-46c7-a711-f26320370541" path="/var/lib/kubelet/pods/a33c345f-a76e-46c7-a711-f26320370541/volumes" Dec 05 12:32:48 crc kubenswrapper[4807]: I1205 12:32:48.158631 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:49 crc kubenswrapper[4807]: I1205 12:32:49.234130 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.466761 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.466847 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.466909 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.467714 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.468098 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" gracePeriod=600 Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.914996 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" exitCode=0 Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.915045 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e"} Dec 05 12:32:52 crc kubenswrapper[4807]: I1205 12:32:52.915075 4807 scope.go:117] "RemoveContainer" containerID="ba4c03761578e8e0b9faf772d0c622df752a9d6198091d0d8315216f608b690f" Dec 05 12:32:53 crc kubenswrapper[4807]: E1205 12:32:53.091251 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:32:53 crc kubenswrapper[4807]: I1205 12:32:53.935492 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:32:53 crc kubenswrapper[4807]: E1205 12:32:53.935805 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:32:54 crc kubenswrapper[4807]: I1205 12:32:54.234464 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:54 crc kubenswrapper[4807]: I1205 12:32:54.240182 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:32:54 crc kubenswrapper[4807]: I1205 12:32:54.948239 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:04 crc kubenswrapper[4807]: I1205 12:33:04.222858 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/keystone-5c74bb55d6-5dfz2" Dec 05 12:33:04 crc kubenswrapper[4807]: I1205 12:33:04.283680 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-694f96b6d4-c2bm4"] Dec 05 12:33:04 crc kubenswrapper[4807]: I1205 12:33:04.283954 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" podUID="50f5f9c8-ef02-49d3-9b04-115ee3f34e93" containerName="keystone-api" containerID="cri-o://b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016" gracePeriod=30 Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.485155 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578333 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-config-data\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578370 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-internal-tls-certs\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578427 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-credential-keys\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578563 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-scripts\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578602 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrjkm\" (UniqueName: \"kubernetes.io/projected/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-kube-api-access-lrjkm\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578616 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-combined-ca-bundle\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578641 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-fernet-keys\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.578661 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-public-tls-certs\") pod \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\" (UID: \"50f5f9c8-ef02-49d3-9b04-115ee3f34e93\") " Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.584208 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.584209 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-kube-api-access-lrjkm" (OuterVolumeSpecName: "kube-api-access-lrjkm") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "kube-api-access-lrjkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.589477 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-scripts" (OuterVolumeSpecName: "scripts") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.594795 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.603733 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.605970 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-config-data" (OuterVolumeSpecName: "config-data") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.628432 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.629445 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "50f5f9c8-ef02-49d3-9b04-115ee3f34e93" (UID: "50f5f9c8-ef02-49d3-9b04-115ee3f34e93"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680519 4807 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680564 4807 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680575 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680583 4807 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680591 4807 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680601 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680609 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrjkm\" (UniqueName: \"kubernetes.io/projected/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-kube-api-access-lrjkm\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.680617 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f5f9c8-ef02-49d3-9b04-115ee3f34e93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.839305 4807 generic.go:334] "Generic (PLEG): container finished" podID="50f5f9c8-ef02-49d3-9b04-115ee3f34e93" containerID="b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016" exitCode=0 Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.839370 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" event={"ID":"50f5f9c8-ef02-49d3-9b04-115ee3f34e93","Type":"ContainerDied","Data":"b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016"} Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.839421 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" event={"ID":"50f5f9c8-ef02-49d3-9b04-115ee3f34e93","Type":"ContainerDied","Data":"6310e9c6db01fcb9092f9b320d76eb7e07b176147418b2077b0b9dd4d9e7832d"} Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.839447 4807 scope.go:117] "RemoveContainer" containerID="b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.839438 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-694f96b6d4-c2bm4" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.875782 4807 scope.go:117] "RemoveContainer" containerID="b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016" Dec 05 12:33:08 crc kubenswrapper[4807]: E1205 12:33:08.878336 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016\": container with ID starting with b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016 not found: ID does not exist" containerID="b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.878552 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016"} err="failed to get container status \"b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016\": rpc error: code = NotFound desc = could not find container \"b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016\": container with ID starting with b3e49886ef86dd0cde5476045708e561cb59d5c868ee28337e4f47fe7fe07016 not found: ID does not exist" Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.926495 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-694f96b6d4-c2bm4"] Dec 05 12:33:08 crc kubenswrapper[4807]: I1205 12:33:08.970550 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-694f96b6d4-c2bm4"] Dec 05 12:33:09 crc kubenswrapper[4807]: I1205 12:33:09.235667 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:33:09 crc kubenswrapper[4807]: E1205 12:33:09.235945 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:33:09 crc kubenswrapper[4807]: I1205 12:33:09.264078 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50f5f9c8-ef02-49d3-9b04-115ee3f34e93" path="/var/lib/kubelet/pods/50f5f9c8-ef02-49d3-9b04-115ee3f34e93/volumes" Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.049031 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.049960 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-central-agent" containerID="cri-o://a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a" gracePeriod=30 Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.050048 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="sg-core" containerID="cri-o://69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495" gracePeriod=30 Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.050072 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-notification-agent" containerID="cri-o://78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65" gracePeriod=30 Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.050395 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="proxy-httpd" containerID="cri-o://f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42" gracePeriod=30 Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.862359 4807 generic.go:334] "Generic (PLEG): container finished" podID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerID="f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42" exitCode=0 Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.862401 4807 generic.go:334] "Generic (PLEG): container finished" podID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerID="69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495" exitCode=2 Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.862411 4807 generic.go:334] "Generic (PLEG): container finished" podID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerID="a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a" exitCode=0 Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.862438 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerDied","Data":"f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42"} Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.862467 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerDied","Data":"69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495"} Dec 05 12:33:10 crc kubenswrapper[4807]: I1205 12:33:10.862482 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerDied","Data":"a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a"} Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.526109 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625655 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-ceilometer-tls-certs\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625728 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbv5g\" (UniqueName: \"kubernetes.io/projected/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-kube-api-access-rbv5g\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625795 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-run-httpd\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625833 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-combined-ca-bundle\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625897 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-scripts\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625922 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-log-httpd\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625951 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-config-data\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.625988 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-sg-core-conf-yaml\") pod \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\" (UID: \"eb7f50a8-b2a9-4a77-92b5-a67a1346c140\") " Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.626226 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.626435 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.626835 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.649734 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-kube-api-access-rbv5g" (OuterVolumeSpecName: "kube-api-access-rbv5g") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "kube-api-access-rbv5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.649741 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-scripts" (OuterVolumeSpecName: "scripts") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.653090 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.710975 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.728874 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.729482 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.729539 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.729551 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.729564 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.729576 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.729588 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbv5g\" (UniqueName: \"kubernetes.io/projected/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-kube-api-access-rbv5g\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.748774 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-config-data" (OuterVolumeSpecName: "config-data") pod "eb7f50a8-b2a9-4a77-92b5-a67a1346c140" (UID: "eb7f50a8-b2a9-4a77-92b5-a67a1346c140"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.830769 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7f50a8-b2a9-4a77-92b5-a67a1346c140-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.904844 4807 generic.go:334] "Generic (PLEG): container finished" podID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerID="78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65" exitCode=0 Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.904892 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerDied","Data":"78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65"} Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.904933 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"eb7f50a8-b2a9-4a77-92b5-a67a1346c140","Type":"ContainerDied","Data":"55ac9b9c6f5bac726ba0ca2e7f8353c8c25127a378e5eaafbb3c749ced9c9714"} Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.904956 4807 scope.go:117] "RemoveContainer" containerID="f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.905218 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.927625 4807 scope.go:117] "RemoveContainer" containerID="69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.939374 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.948319 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.950265 4807 scope.go:117] "RemoveContainer" containerID="78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.965615 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.965943 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f5f9c8-ef02-49d3-9b04-115ee3f34e93" containerName="keystone-api" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.965956 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f5f9c8-ef02-49d3-9b04-115ee3f34e93" containerName="keystone-api" Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.965969 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="extract-content" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.965975 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="extract-content" Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.965985 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="proxy-httpd" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.965992 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="proxy-httpd" Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.966006 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-central-agent" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966012 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-central-agent" Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.966020 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="extract-utilities" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966025 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="extract-utilities" Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.966034 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-notification-agent" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966040 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-notification-agent" Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.966056 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="sg-core" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966062 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="sg-core" Dec 05 12:33:15 crc kubenswrapper[4807]: E1205 12:33:15.966071 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="registry-server" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966076 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="registry-server" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966226 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="sg-core" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966242 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-central-agent" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966252 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33c345f-a76e-46c7-a711-f26320370541" containerName="registry-server" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966262 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="50f5f9c8-ef02-49d3-9b04-115ee3f34e93" containerName="keystone-api" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966271 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="proxy-httpd" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.966283 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" containerName="ceilometer-notification-agent" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.967833 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.971317 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.971728 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.977254 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:33:15 crc kubenswrapper[4807]: I1205 12:33:15.990944 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.001191 4807 scope.go:117] "RemoveContainer" containerID="a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.049893 4807 scope.go:117] "RemoveContainer" containerID="f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42" Dec 05 12:33:16 crc kubenswrapper[4807]: E1205 12:33:16.051134 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42\": container with ID starting with f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42 not found: ID does not exist" containerID="f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.051180 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42"} err="failed to get container status \"f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42\": rpc error: code = NotFound desc = could not find container \"f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42\": container with ID starting with f009e4a39dbfac165570b28cb8e33adbf37d82c000a307dffcadc3645b008e42 not found: ID does not exist" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.051211 4807 scope.go:117] "RemoveContainer" containerID="69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495" Dec 05 12:33:16 crc kubenswrapper[4807]: E1205 12:33:16.054264 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495\": container with ID starting with 69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495 not found: ID does not exist" containerID="69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.054309 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495"} err="failed to get container status \"69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495\": rpc error: code = NotFound desc = could not find container \"69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495\": container with ID starting with 69121056d60ec2e5e631ad98cae976ea96a9c2c9fd22917278feb6571368b495 not found: ID does not exist" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.054339 4807 scope.go:117] "RemoveContainer" containerID="78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65" Dec 05 12:33:16 crc kubenswrapper[4807]: E1205 12:33:16.055660 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65\": container with ID starting with 78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65 not found: ID does not exist" containerID="78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.055697 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65"} err="failed to get container status \"78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65\": rpc error: code = NotFound desc = could not find container \"78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65\": container with ID starting with 78e226db9a2aedb92c0beeae4725b585e5cc70b6301f243ec28807d48b150f65 not found: ID does not exist" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.056591 4807 scope.go:117] "RemoveContainer" containerID="a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a" Dec 05 12:33:16 crc kubenswrapper[4807]: E1205 12:33:16.056941 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a\": container with ID starting with a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a not found: ID does not exist" containerID="a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.056978 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a"} err="failed to get container status \"a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a\": rpc error: code = NotFound desc = could not find container \"a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a\": container with ID starting with a7fd88ae2d015eac2df8cf64a479d3d4ccade007303568f2e440dee58fcce95a not found: ID does not exist" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137200 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-scripts\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137333 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137361 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137434 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-config-data\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137502 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-run-httpd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137651 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsngd\" (UniqueName: \"kubernetes.io/projected/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-kube-api-access-gsngd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137732 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.137870 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-log-httpd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.239702 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.239744 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.239782 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-config-data\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.239811 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-run-httpd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.239876 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsngd\" (UniqueName: \"kubernetes.io/projected/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-kube-api-access-gsngd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.239924 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.239982 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-log-httpd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.240033 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-scripts\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.241971 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-run-httpd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.242010 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-log-httpd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.245545 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.245655 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.246771 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-config-data\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.248079 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.253262 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-scripts\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.258495 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsngd\" (UniqueName: \"kubernetes.io/projected/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-kube-api-access-gsngd\") pod \"ceilometer-0\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.333230 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.847359 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:16 crc kubenswrapper[4807]: I1205 12:33:16.918031 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerStarted","Data":"cd406990f7e342ef7d25a971c65a6625d2bc0e55980238e5dbe0a0c27d1b53df"} Dec 05 12:33:17 crc kubenswrapper[4807]: I1205 12:33:17.255940 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb7f50a8-b2a9-4a77-92b5-a67a1346c140" path="/var/lib/kubelet/pods/eb7f50a8-b2a9-4a77-92b5-a67a1346c140/volumes" Dec 05 12:33:17 crc kubenswrapper[4807]: I1205 12:33:17.928445 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerStarted","Data":"02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f"} Dec 05 12:33:18 crc kubenswrapper[4807]: I1205 12:33:18.938039 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerStarted","Data":"a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b"} Dec 05 12:33:19 crc kubenswrapper[4807]: I1205 12:33:19.947428 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerStarted","Data":"d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13"} Dec 05 12:33:20 crc kubenswrapper[4807]: I1205 12:33:20.956636 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerStarted","Data":"87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53"} Dec 05 12:33:20 crc kubenswrapper[4807]: I1205 12:33:20.957898 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:20 crc kubenswrapper[4807]: I1205 12:33:20.986058 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.455662819 podStartE2EDuration="5.986041338s" podCreationTimestamp="2025-12-05 12:33:15 +0000 UTC" firstStartedPulling="2025-12-05 12:33:16.851820039 +0000 UTC m=+1626.345683308" lastFinishedPulling="2025-12-05 12:33:20.382198558 +0000 UTC m=+1629.876061827" observedRunningTime="2025-12-05 12:33:20.977138311 +0000 UTC m=+1630.471001580" watchObservedRunningTime="2025-12-05 12:33:20.986041338 +0000 UTC m=+1630.479904607" Dec 05 12:33:23 crc kubenswrapper[4807]: I1205 12:33:23.237005 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:33:23 crc kubenswrapper[4807]: E1205 12:33:23.238006 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:33:36 crc kubenswrapper[4807]: I1205 12:33:36.235848 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:33:36 crc kubenswrapper[4807]: E1205 12:33:36.236561 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.275146 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh"] Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.283516 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-6s9wh"] Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.329066 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.329353 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="58a7640d-9cfa-4e9b-9501-5e2b4da9a886" containerName="watcher-applier" containerID="cri-o://51dde105a0fac8edef020314a4311b0a44cac5609616e2b8f5f288192934fb9e" gracePeriod=30 Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.366545 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcherdf4f-account-delete-rtqpc"] Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.367877 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.386766 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherdf4f-account-delete-rtqpc"] Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.406474 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.412612 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-kuttl-api-log" containerID="cri-o://2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6" gracePeriod=30 Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.412770 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-api" containerID="cri-o://3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92" gracePeriod=30 Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.437299 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c07f685-acbb-4829-8fa4-3767be80b229-operator-scripts\") pod \"watcherdf4f-account-delete-rtqpc\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.437408 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnqdd\" (UniqueName: \"kubernetes.io/projected/1c07f685-acbb-4829-8fa4-3767be80b229-kube-api-access-mnqdd\") pod \"watcherdf4f-account-delete-rtqpc\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.452030 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.452281 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="8e820190-88c8-4df3-9832-2d9d0bb914cc" containerName="watcher-decision-engine" containerID="cri-o://a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba" gracePeriod=30 Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.541295 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnqdd\" (UniqueName: \"kubernetes.io/projected/1c07f685-acbb-4829-8fa4-3767be80b229-kube-api-access-mnqdd\") pod \"watcherdf4f-account-delete-rtqpc\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.541380 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c07f685-acbb-4829-8fa4-3767be80b229-operator-scripts\") pod \"watcherdf4f-account-delete-rtqpc\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.542131 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c07f685-acbb-4829-8fa4-3767be80b229-operator-scripts\") pod \"watcherdf4f-account-delete-rtqpc\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.561576 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnqdd\" (UniqueName: \"kubernetes.io/projected/1c07f685-acbb-4829-8fa4-3767be80b229-kube-api-access-mnqdd\") pod \"watcherdf4f-account-delete-rtqpc\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:40 crc kubenswrapper[4807]: I1205 12:33:40.699093 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:41 crc kubenswrapper[4807]: I1205 12:33:41.129933 4807 generic.go:334] "Generic (PLEG): container finished" podID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerID="2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6" exitCode=143 Dec 05 12:33:41 crc kubenswrapper[4807]: I1205 12:33:41.130139 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"98619dc2-c4b1-45e5-8c69-a8410aafa29a","Type":"ContainerDied","Data":"2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6"} Dec 05 12:33:41 crc kubenswrapper[4807]: I1205 12:33:41.243577 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6c66668-1556-427e-be1d-1d36a93e1369" path="/var/lib/kubelet/pods/c6c66668-1556-427e-be1d-1d36a93e1369/volumes" Dec 05 12:33:41 crc kubenswrapper[4807]: I1205 12:33:41.247877 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherdf4f-account-delete-rtqpc"] Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.156030 4807 generic.go:334] "Generic (PLEG): container finished" podID="1c07f685-acbb-4829-8fa4-3767be80b229" containerID="d8f0f5ffc9b8753c94122ddf3f3ac4650009d358ffe8ead4664c118401ee1833" exitCode=0 Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.156402 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" event={"ID":"1c07f685-acbb-4829-8fa4-3767be80b229","Type":"ContainerDied","Data":"d8f0f5ffc9b8753c94122ddf3f3ac4650009d358ffe8ead4664c118401ee1833"} Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.156448 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" event={"ID":"1c07f685-acbb-4829-8fa4-3767be80b229","Type":"ContainerStarted","Data":"7394ea7054e4b316c8b445136d1cdcfcb33ca7aab166a0438e26da22f5b5b17b"} Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.745559 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.887864 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlzxw\" (UniqueName: \"kubernetes.io/projected/98619dc2-c4b1-45e5-8c69-a8410aafa29a-kube-api-access-nlzxw\") pod \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.887981 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-cert-memcached-mtls\") pod \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.887999 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-custom-prometheus-ca\") pod \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.888072 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98619dc2-c4b1-45e5-8c69-a8410aafa29a-logs\") pod \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.888141 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-combined-ca-bundle\") pod \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.888182 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-config-data\") pod \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\" (UID: \"98619dc2-c4b1-45e5-8c69-a8410aafa29a\") " Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.888640 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98619dc2-c4b1-45e5-8c69-a8410aafa29a-logs" (OuterVolumeSpecName: "logs") pod "98619dc2-c4b1-45e5-8c69-a8410aafa29a" (UID: "98619dc2-c4b1-45e5-8c69-a8410aafa29a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.892991 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98619dc2-c4b1-45e5-8c69-a8410aafa29a-kube-api-access-nlzxw" (OuterVolumeSpecName: "kube-api-access-nlzxw") pod "98619dc2-c4b1-45e5-8c69-a8410aafa29a" (UID: "98619dc2-c4b1-45e5-8c69-a8410aafa29a"). InnerVolumeSpecName "kube-api-access-nlzxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.917830 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98619dc2-c4b1-45e5-8c69-a8410aafa29a" (UID: "98619dc2-c4b1-45e5-8c69-a8410aafa29a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.922351 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "98619dc2-c4b1-45e5-8c69-a8410aafa29a" (UID: "98619dc2-c4b1-45e5-8c69-a8410aafa29a"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.938783 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-config-data" (OuterVolumeSpecName: "config-data") pod "98619dc2-c4b1-45e5-8c69-a8410aafa29a" (UID: "98619dc2-c4b1-45e5-8c69-a8410aafa29a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.978395 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "98619dc2-c4b1-45e5-8c69-a8410aafa29a" (UID: "98619dc2-c4b1-45e5-8c69-a8410aafa29a"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.990320 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98619dc2-c4b1-45e5-8c69-a8410aafa29a-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.990361 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.990377 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.990390 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlzxw\" (UniqueName: \"kubernetes.io/projected/98619dc2-c4b1-45e5-8c69-a8410aafa29a-kube-api-access-nlzxw\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.990403 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:42 crc kubenswrapper[4807]: I1205 12:33:42.990414 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/98619dc2-c4b1-45e5-8c69-a8410aafa29a-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.167203 4807 generic.go:334] "Generic (PLEG): container finished" podID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerID="3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92" exitCode=0 Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.167392 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.173834 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"98619dc2-c4b1-45e5-8c69-a8410aafa29a","Type":"ContainerDied","Data":"3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92"} Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.174192 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"98619dc2-c4b1-45e5-8c69-a8410aafa29a","Type":"ContainerDied","Data":"b998469ef6a72ccec6a784a06a2d00a33c842db4286fa9f6ee281525a57ce573"} Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.174343 4807 scope.go:117] "RemoveContainer" containerID="3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.200603 4807 scope.go:117] "RemoveContainer" containerID="2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.206478 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.214304 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.223005 4807 scope.go:117] "RemoveContainer" containerID="3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92" Dec 05 12:33:43 crc kubenswrapper[4807]: E1205 12:33:43.224717 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92\": container with ID starting with 3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92 not found: ID does not exist" containerID="3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.224756 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92"} err="failed to get container status \"3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92\": rpc error: code = NotFound desc = could not find container \"3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92\": container with ID starting with 3cfe016a25b7172d07e33f52c23e9638e7cde8d92ea24a4c002f7e3458b49b92 not found: ID does not exist" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.224786 4807 scope.go:117] "RemoveContainer" containerID="2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6" Dec 05 12:33:43 crc kubenswrapper[4807]: E1205 12:33:43.225026 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6\": container with ID starting with 2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6 not found: ID does not exist" containerID="2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.225053 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6"} err="failed to get container status \"2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6\": rpc error: code = NotFound desc = could not find container \"2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6\": container with ID starting with 2d880b1d803d39213ee21fbc8c781bf2d240a7fec16042caf9bf5b2e328178f6 not found: ID does not exist" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.247437 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" path="/var/lib/kubelet/pods/98619dc2-c4b1-45e5-8c69-a8410aafa29a/volumes" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.474175 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.604762 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnqdd\" (UniqueName: \"kubernetes.io/projected/1c07f685-acbb-4829-8fa4-3767be80b229-kube-api-access-mnqdd\") pod \"1c07f685-acbb-4829-8fa4-3767be80b229\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.604925 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c07f685-acbb-4829-8fa4-3767be80b229-operator-scripts\") pod \"1c07f685-acbb-4829-8fa4-3767be80b229\" (UID: \"1c07f685-acbb-4829-8fa4-3767be80b229\") " Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.605569 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c07f685-acbb-4829-8fa4-3767be80b229-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c07f685-acbb-4829-8fa4-3767be80b229" (UID: "1c07f685-acbb-4829-8fa4-3767be80b229"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.607903 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c07f685-acbb-4829-8fa4-3767be80b229-kube-api-access-mnqdd" (OuterVolumeSpecName: "kube-api-access-mnqdd") pod "1c07f685-acbb-4829-8fa4-3767be80b229" (UID: "1c07f685-acbb-4829-8fa4-3767be80b229"). InnerVolumeSpecName "kube-api-access-mnqdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.707027 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c07f685-acbb-4829-8fa4-3767be80b229-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:43 crc kubenswrapper[4807]: I1205 12:33:43.707076 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnqdd\" (UniqueName: \"kubernetes.io/projected/1c07f685-acbb-4829-8fa4-3767be80b229-kube-api-access-mnqdd\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.101136 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.101417 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-central-agent" containerID="cri-o://02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f" gracePeriod=30 Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.103708 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="proxy-httpd" containerID="cri-o://87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53" gracePeriod=30 Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.103814 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="sg-core" containerID="cri-o://d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13" gracePeriod=30 Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.103853 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-notification-agent" containerID="cri-o://a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b" gracePeriod=30 Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.113672 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.180026 4807 generic.go:334] "Generic (PLEG): container finished" podID="58a7640d-9cfa-4e9b-9501-5e2b4da9a886" containerID="51dde105a0fac8edef020314a4311b0a44cac5609616e2b8f5f288192934fb9e" exitCode=0 Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.180095 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"58a7640d-9cfa-4e9b-9501-5e2b4da9a886","Type":"ContainerDied","Data":"51dde105a0fac8edef020314a4311b0a44cac5609616e2b8f5f288192934fb9e"} Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.182773 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" event={"ID":"1c07f685-acbb-4829-8fa4-3767be80b229","Type":"ContainerDied","Data":"7394ea7054e4b316c8b445136d1cdcfcb33ca7aab166a0438e26da22f5b5b17b"} Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.182822 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7394ea7054e4b316c8b445136d1cdcfcb33ca7aab166a0438e26da22f5b5b17b" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.182822 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherdf4f-account-delete-rtqpc" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.471226 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.622222 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnxh7\" (UniqueName: \"kubernetes.io/projected/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-kube-api-access-vnxh7\") pod \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.622400 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-config-data\") pod \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.622455 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-logs\") pod \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.622561 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-combined-ca-bundle\") pod \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.622692 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-cert-memcached-mtls\") pod \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\" (UID: \"58a7640d-9cfa-4e9b-9501-5e2b4da9a886\") " Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.624254 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-logs" (OuterVolumeSpecName: "logs") pod "58a7640d-9cfa-4e9b-9501-5e2b4da9a886" (UID: "58a7640d-9cfa-4e9b-9501-5e2b4da9a886"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.640815 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-kube-api-access-vnxh7" (OuterVolumeSpecName: "kube-api-access-vnxh7") pod "58a7640d-9cfa-4e9b-9501-5e2b4da9a886" (UID: "58a7640d-9cfa-4e9b-9501-5e2b4da9a886"). InnerVolumeSpecName "kube-api-access-vnxh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.657814 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58a7640d-9cfa-4e9b-9501-5e2b4da9a886" (UID: "58a7640d-9cfa-4e9b-9501-5e2b4da9a886"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.720664 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-config-data" (OuterVolumeSpecName: "config-data") pod "58a7640d-9cfa-4e9b-9501-5e2b4da9a886" (UID: "58a7640d-9cfa-4e9b-9501-5e2b4da9a886"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.733052 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.733084 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.733097 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnxh7\" (UniqueName: \"kubernetes.io/projected/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-kube-api-access-vnxh7\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.733109 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.741149 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "58a7640d-9cfa-4e9b-9501-5e2b4da9a886" (UID: "58a7640d-9cfa-4e9b-9501-5e2b4da9a886"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:44 crc kubenswrapper[4807]: I1205 12:33:44.834765 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58a7640d-9cfa-4e9b-9501-5e2b4da9a886-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.193255 4807 generic.go:334] "Generic (PLEG): container finished" podID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerID="87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53" exitCode=0 Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.193292 4807 generic.go:334] "Generic (PLEG): container finished" podID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerID="d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13" exitCode=2 Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.193303 4807 generic.go:334] "Generic (PLEG): container finished" podID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerID="02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f" exitCode=0 Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.193326 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerDied","Data":"87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53"} Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.193365 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerDied","Data":"d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13"} Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.193376 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerDied","Data":"02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f"} Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.195648 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"58a7640d-9cfa-4e9b-9501-5e2b4da9a886","Type":"ContainerDied","Data":"b6c1b11f5290a9537397ea13e8d5d426ec87efed8952995dfbf3efd6984ffa57"} Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.195707 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.195712 4807 scope.go:117] "RemoveContainer" containerID="51dde105a0fac8edef020314a4311b0a44cac5609616e2b8f5f288192934fb9e" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.294632 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.306736 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.421601 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-mw4nj"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.426206 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-mw4nj"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.509487 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.531783 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcherdf4f-account-delete-rtqpc"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.538458 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-df4f-account-create-update-2g9k2"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.544879 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcherdf4f-account-delete-rtqpc"] Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.600579 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.762935 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-ceilometer-tls-certs\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.762999 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-scripts\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763048 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-run-httpd\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763106 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-log-httpd\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763138 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-combined-ca-bundle\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763205 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsngd\" (UniqueName: \"kubernetes.io/projected/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-kube-api-access-gsngd\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763253 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-sg-core-conf-yaml\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763276 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-config-data\") pod \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\" (UID: \"47e7c8e3-ecca-4098-a792-c07fdb8f23f7\") " Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763379 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763583 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.763728 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.775373 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-scripts" (OuterVolumeSpecName: "scripts") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.777726 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-kube-api-access-gsngd" (OuterVolumeSpecName: "kube-api-access-gsngd") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "kube-api-access-gsngd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.799733 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.865239 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.865268 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.865277 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsngd\" (UniqueName: \"kubernetes.io/projected/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-kube-api-access-gsngd\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.865286 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.873682 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.873702 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.899670 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-config-data" (OuterVolumeSpecName: "config-data") pod "47e7c8e3-ecca-4098-a792-c07fdb8f23f7" (UID: "47e7c8e3-ecca-4098-a792-c07fdb8f23f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.966696 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.966736 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:45 crc kubenswrapper[4807]: I1205 12:33:45.966746 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47e7c8e3-ecca-4098-a792-c07fdb8f23f7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.156213 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.213319 4807 generic.go:334] "Generic (PLEG): container finished" podID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerID="a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b" exitCode=0 Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.213405 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerDied","Data":"a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b"} Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.213488 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"47e7c8e3-ecca-4098-a792-c07fdb8f23f7","Type":"ContainerDied","Data":"cd406990f7e342ef7d25a971c65a6625d2bc0e55980238e5dbe0a0c27d1b53df"} Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.213546 4807 scope.go:117] "RemoveContainer" containerID="87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.213741 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.216638 4807 generic.go:334] "Generic (PLEG): container finished" podID="8e820190-88c8-4df3-9832-2d9d0bb914cc" containerID="a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba" exitCode=0 Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.216669 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"8e820190-88c8-4df3-9832-2d9d0bb914cc","Type":"ContainerDied","Data":"a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba"} Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.216690 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"8e820190-88c8-4df3-9832-2d9d0bb914cc","Type":"ContainerDied","Data":"03ac910a1a94f42c3fcd4d63423d15258385fe87d9fda90087425a8a2802a161"} Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.216742 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.259354 4807 scope.go:117] "RemoveContainer" containerID="d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.263682 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.272640 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e820190-88c8-4df3-9832-2d9d0bb914cc-logs\") pod \"8e820190-88c8-4df3-9832-2d9d0bb914cc\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.272697 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpkw5\" (UniqueName: \"kubernetes.io/projected/8e820190-88c8-4df3-9832-2d9d0bb914cc-kube-api-access-mpkw5\") pod \"8e820190-88c8-4df3-9832-2d9d0bb914cc\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.272780 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-cert-memcached-mtls\") pod \"8e820190-88c8-4df3-9832-2d9d0bb914cc\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.272960 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-custom-prometheus-ca\") pod \"8e820190-88c8-4df3-9832-2d9d0bb914cc\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.273014 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-combined-ca-bundle\") pod \"8e820190-88c8-4df3-9832-2d9d0bb914cc\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.273037 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-config-data\") pod \"8e820190-88c8-4df3-9832-2d9d0bb914cc\" (UID: \"8e820190-88c8-4df3-9832-2d9d0bb914cc\") " Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.276118 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e820190-88c8-4df3-9832-2d9d0bb914cc-logs" (OuterVolumeSpecName: "logs") pod "8e820190-88c8-4df3-9832-2d9d0bb914cc" (UID: "8e820190-88c8-4df3-9832-2d9d0bb914cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.277206 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.287745 4807 scope.go:117] "RemoveContainer" containerID="a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.293654 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e820190-88c8-4df3-9832-2d9d0bb914cc-kube-api-access-mpkw5" (OuterVolumeSpecName: "kube-api-access-mpkw5") pod "8e820190-88c8-4df3-9832-2d9d0bb914cc" (UID: "8e820190-88c8-4df3-9832-2d9d0bb914cc"). InnerVolumeSpecName "kube-api-access-mpkw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.311835 4807 scope.go:117] "RemoveContainer" containerID="02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.335814 4807 scope.go:117] "RemoveContainer" containerID="87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.336287 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53\": container with ID starting with 87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53 not found: ID does not exist" containerID="87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.336353 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53"} err="failed to get container status \"87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53\": rpc error: code = NotFound desc = could not find container \"87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53\": container with ID starting with 87c55a23a62d51e7999e50e120a0bec4561d9b5f8e8004b6bf2481cd24499f53 not found: ID does not exist" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.336376 4807 scope.go:117] "RemoveContainer" containerID="d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.340751 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13\": container with ID starting with d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13 not found: ID does not exist" containerID="d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.340810 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13"} err="failed to get container status \"d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13\": rpc error: code = NotFound desc = could not find container \"d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13\": container with ID starting with d74d0c6a7d9241d6eb1fb5eaa938d9865487ebf429b0f88703df37b30f999e13 not found: ID does not exist" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.340845 4807 scope.go:117] "RemoveContainer" containerID="a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.341243 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b\": container with ID starting with a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b not found: ID does not exist" containerID="a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.341272 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b"} err="failed to get container status \"a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b\": rpc error: code = NotFound desc = could not find container \"a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b\": container with ID starting with a4752ac5aa5d257306901a894770a29d9b6a5c564f24ca6a4ad9150c8f57f47b not found: ID does not exist" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.341285 4807 scope.go:117] "RemoveContainer" containerID="02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.341686 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f\": container with ID starting with 02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f not found: ID does not exist" containerID="02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.341715 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f"} err="failed to get container status \"02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f\": rpc error: code = NotFound desc = could not find container \"02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f\": container with ID starting with 02d886e51f1f4b02faf7d4c54af06e6172c4e6091d8169cca1724aa88e45bc4f not found: ID does not exist" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.341730 4807 scope.go:117] "RemoveContainer" containerID="a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346037 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346476 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-notification-agent" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346491 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-notification-agent" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346505 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a7640d-9cfa-4e9b-9501-5e2b4da9a886" containerName="watcher-applier" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346513 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a7640d-9cfa-4e9b-9501-5e2b4da9a886" containerName="watcher-applier" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346544 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-kuttl-api-log" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346554 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-kuttl-api-log" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346571 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c07f685-acbb-4829-8fa4-3767be80b229" containerName="mariadb-account-delete" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346580 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c07f685-acbb-4829-8fa4-3767be80b229" containerName="mariadb-account-delete" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346592 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="sg-core" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346599 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="sg-core" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346616 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e820190-88c8-4df3-9832-2d9d0bb914cc" containerName="watcher-decision-engine" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346627 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e820190-88c8-4df3-9832-2d9d0bb914cc" containerName="watcher-decision-engine" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346647 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-api" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346657 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-api" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346668 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="proxy-httpd" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346781 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="proxy-httpd" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.346798 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-central-agent" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346806 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-central-agent" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346977 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-kuttl-api-log" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.346995 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="98619dc2-c4b1-45e5-8c69-a8410aafa29a" containerName="watcher-api" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.347004 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a7640d-9cfa-4e9b-9501-5e2b4da9a886" containerName="watcher-applier" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.347014 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-notification-agent" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.347022 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="proxy-httpd" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.347037 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e820190-88c8-4df3-9832-2d9d0bb914cc" containerName="watcher-decision-engine" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.347050 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="ceilometer-central-agent" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.347061 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" containerName="sg-core" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.347078 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c07f685-acbb-4829-8fa4-3767be80b229" containerName="mariadb-account-delete" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.348848 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.362873 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.364495 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.365245 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.365333 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.374544 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e820190-88c8-4df3-9832-2d9d0bb914cc-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.374576 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpkw5\" (UniqueName: \"kubernetes.io/projected/8e820190-88c8-4df3-9832-2d9d0bb914cc-kube-api-access-mpkw5\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.380809 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "8e820190-88c8-4df3-9832-2d9d0bb914cc" (UID: "8e820190-88c8-4df3-9832-2d9d0bb914cc"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.381665 4807 scope.go:117] "RemoveContainer" containerID="a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.386603 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e820190-88c8-4df3-9832-2d9d0bb914cc" (UID: "8e820190-88c8-4df3-9832-2d9d0bb914cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:46 crc kubenswrapper[4807]: E1205 12:33:46.386931 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba\": container with ID starting with a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba not found: ID does not exist" containerID="a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.386986 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba"} err="failed to get container status \"a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba\": rpc error: code = NotFound desc = could not find container \"a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba\": container with ID starting with a1bf893e623fa00acf60110222d4ee69a8d257e93aa39632a37e86adc7648dba not found: ID does not exist" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.396753 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-config-data" (OuterVolumeSpecName: "config-data") pod "8e820190-88c8-4df3-9832-2d9d0bb914cc" (UID: "8e820190-88c8-4df3-9832-2d9d0bb914cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.435675 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "8e820190-88c8-4df3-9832-2d9d0bb914cc" (UID: "8e820190-88c8-4df3-9832-2d9d0bb914cc"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476433 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476498 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-scripts\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476542 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-log-httpd\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476651 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-run-httpd\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476737 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476808 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-config-data\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476889 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh9kr\" (UniqueName: \"kubernetes.io/projected/ddf023d3-0bd1-4a7d-a622-3442787a0309-kube-api-access-qh9kr\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.476960 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.477054 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.477076 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.477088 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.477099 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8e820190-88c8-4df3-9832-2d9d0bb914cc-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.545023 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.551715 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578360 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578430 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-scripts\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578462 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-log-httpd\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578506 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-run-httpd\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578561 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578591 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-config-data\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578638 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh9kr\" (UniqueName: \"kubernetes.io/projected/ddf023d3-0bd1-4a7d-a622-3442787a0309-kube-api-access-qh9kr\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.578681 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.579649 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-run-httpd\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.579989 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-log-httpd\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.583144 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.583176 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.583613 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-config-data\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.584618 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-scripts\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.587231 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.609436 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh9kr\" (UniqueName: \"kubernetes.io/projected/ddf023d3-0bd1-4a7d-a622-3442787a0309-kube-api-access-qh9kr\") pod \"ceilometer-0\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:46 crc kubenswrapper[4807]: I1205 12:33:46.681386 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.153857 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.157832 4807 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.227419 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerStarted","Data":"5b3f5debe10a56df48378cf8df77d0a0ba580c9876fce8f7b103b79428cd3f7c"} Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.235603 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:33:47 crc kubenswrapper[4807]: E1205 12:33:47.235897 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.244090 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c07f685-acbb-4829-8fa4-3767be80b229" path="/var/lib/kubelet/pods/1c07f685-acbb-4829-8fa4-3767be80b229/volumes" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.244764 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47e7c8e3-ecca-4098-a792-c07fdb8f23f7" path="/var/lib/kubelet/pods/47e7c8e3-ecca-4098-a792-c07fdb8f23f7/volumes" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.245744 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58a7640d-9cfa-4e9b-9501-5e2b4da9a886" path="/var/lib/kubelet/pods/58a7640d-9cfa-4e9b-9501-5e2b4da9a886/volumes" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.246683 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a0cd735-40cf-49aa-be5a-2378115ea996" path="/var/lib/kubelet/pods/6a0cd735-40cf-49aa-be5a-2378115ea996/volumes" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.247188 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e820190-88c8-4df3-9832-2d9d0bb914cc" path="/var/lib/kubelet/pods/8e820190-88c8-4df3-9832-2d9d0bb914cc/volumes" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.247673 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc06c206-269f-48e7-b014-f1a7d9f596ac" path="/var/lib/kubelet/pods/cc06c206-269f-48e7-b014-f1a7d9f596ac/volumes" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.489374 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-lgcpq"] Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.490775 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.500494 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lgcpq"] Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.526129 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-4079-account-create-update-fcz88"] Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.527243 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.529924 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.537781 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-4079-account-create-update-fcz88"] Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.593744 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-operator-scripts\") pod \"watcher-db-create-lgcpq\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.593842 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckxn7\" (UniqueName: \"kubernetes.io/projected/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-kube-api-access-ckxn7\") pod \"watcher-db-create-lgcpq\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.695703 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-operator-scripts\") pod \"watcher-db-create-lgcpq\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.695885 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsc8l\" (UniqueName: \"kubernetes.io/projected/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-kube-api-access-dsc8l\") pod \"watcher-4079-account-create-update-fcz88\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.695983 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckxn7\" (UniqueName: \"kubernetes.io/projected/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-kube-api-access-ckxn7\") pod \"watcher-db-create-lgcpq\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.696068 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-operator-scripts\") pod \"watcher-4079-account-create-update-fcz88\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.696478 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-operator-scripts\") pod \"watcher-db-create-lgcpq\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.712572 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckxn7\" (UniqueName: \"kubernetes.io/projected/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-kube-api-access-ckxn7\") pod \"watcher-db-create-lgcpq\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.797808 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsc8l\" (UniqueName: \"kubernetes.io/projected/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-kube-api-access-dsc8l\") pod \"watcher-4079-account-create-update-fcz88\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.797884 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-operator-scripts\") pod \"watcher-4079-account-create-update-fcz88\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.798542 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-operator-scripts\") pod \"watcher-4079-account-create-update-fcz88\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.812372 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.817108 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsc8l\" (UniqueName: \"kubernetes.io/projected/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-kube-api-access-dsc8l\") pod \"watcher-4079-account-create-update-fcz88\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:47 crc kubenswrapper[4807]: I1205 12:33:47.846864 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:48 crc kubenswrapper[4807]: I1205 12:33:48.251462 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerStarted","Data":"5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f"} Dec 05 12:33:48 crc kubenswrapper[4807]: I1205 12:33:48.363803 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-4079-account-create-update-fcz88"] Dec 05 12:33:48 crc kubenswrapper[4807]: I1205 12:33:48.492933 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lgcpq"] Dec 05 12:33:49 crc kubenswrapper[4807]: E1205 12:33:49.076048 4807 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eb80014_c9a5_44e0_92f4_8f768ed9ed25.slice/crio-conmon-60c7c8c7cd2f87db9f4e08ad775231377716b19983a5a65957bcf3909cff53f7.scope\": RecentStats: unable to find data in memory cache]" Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.262358 4807 generic.go:334] "Generic (PLEG): container finished" podID="1819ab6b-5aab-47ee-ac2d-8afad8ec0b36" containerID="178e105ec129899fd78b2dc51dcdbbed1335d1980a83ec82b2fd0c37caf08286" exitCode=0 Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.262432 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" event={"ID":"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36","Type":"ContainerDied","Data":"178e105ec129899fd78b2dc51dcdbbed1335d1980a83ec82b2fd0c37caf08286"} Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.262458 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" event={"ID":"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36","Type":"ContainerStarted","Data":"5210bd9dc76fa5d680255fdd30a12e91b3eef31677325a52e51de6d0493ed77c"} Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.264998 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerStarted","Data":"067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6"} Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.265022 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerStarted","Data":"236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669"} Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.266892 4807 generic.go:334] "Generic (PLEG): container finished" podID="9eb80014-c9a5-44e0-92f4-8f768ed9ed25" containerID="60c7c8c7cd2f87db9f4e08ad775231377716b19983a5a65957bcf3909cff53f7" exitCode=0 Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.266924 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-lgcpq" event={"ID":"9eb80014-c9a5-44e0-92f4-8f768ed9ed25","Type":"ContainerDied","Data":"60c7c8c7cd2f87db9f4e08ad775231377716b19983a5a65957bcf3909cff53f7"} Dec 05 12:33:49 crc kubenswrapper[4807]: I1205 12:33:49.266939 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-lgcpq" event={"ID":"9eb80014-c9a5-44e0-92f4-8f768ed9ed25","Type":"ContainerStarted","Data":"caaa32ad0f1f1219f63b4587a9a5359139d8014785805dbd98f7e5a1507d8b3b"} Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.718402 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.726341 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.849956 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckxn7\" (UniqueName: \"kubernetes.io/projected/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-kube-api-access-ckxn7\") pod \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.850046 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-operator-scripts\") pod \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.850075 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-operator-scripts\") pod \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\" (UID: \"9eb80014-c9a5-44e0-92f4-8f768ed9ed25\") " Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.850238 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsc8l\" (UniqueName: \"kubernetes.io/projected/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-kube-api-access-dsc8l\") pod \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\" (UID: \"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36\") " Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.850874 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1819ab6b-5aab-47ee-ac2d-8afad8ec0b36" (UID: "1819ab6b-5aab-47ee-ac2d-8afad8ec0b36"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.850988 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9eb80014-c9a5-44e0-92f4-8f768ed9ed25" (UID: "9eb80014-c9a5-44e0-92f4-8f768ed9ed25"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.856188 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-kube-api-access-dsc8l" (OuterVolumeSpecName: "kube-api-access-dsc8l") pod "1819ab6b-5aab-47ee-ac2d-8afad8ec0b36" (UID: "1819ab6b-5aab-47ee-ac2d-8afad8ec0b36"). InnerVolumeSpecName "kube-api-access-dsc8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.856254 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-kube-api-access-ckxn7" (OuterVolumeSpecName: "kube-api-access-ckxn7") pod "9eb80014-c9a5-44e0-92f4-8f768ed9ed25" (UID: "9eb80014-c9a5-44e0-92f4-8f768ed9ed25"). InnerVolumeSpecName "kube-api-access-ckxn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.951826 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckxn7\" (UniqueName: \"kubernetes.io/projected/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-kube-api-access-ckxn7\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.951875 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.951892 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb80014-c9a5-44e0-92f4-8f768ed9ed25-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:50 crc kubenswrapper[4807]: I1205 12:33:50.951904 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsc8l\" (UniqueName: \"kubernetes.io/projected/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36-kube-api-access-dsc8l\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.291156 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.291217 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-4079-account-create-update-fcz88" event={"ID":"1819ab6b-5aab-47ee-ac2d-8afad8ec0b36","Type":"ContainerDied","Data":"5210bd9dc76fa5d680255fdd30a12e91b3eef31677325a52e51de6d0493ed77c"} Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.291253 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5210bd9dc76fa5d680255fdd30a12e91b3eef31677325a52e51de6d0493ed77c" Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.294762 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerStarted","Data":"d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6"} Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.296733 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-lgcpq" event={"ID":"9eb80014-c9a5-44e0-92f4-8f768ed9ed25","Type":"ContainerDied","Data":"caaa32ad0f1f1219f63b4587a9a5359139d8014785805dbd98f7e5a1507d8b3b"} Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.296765 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caaa32ad0f1f1219f63b4587a9a5359139d8014785805dbd98f7e5a1507d8b3b" Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.296836 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-lgcpq" Dec 05 12:33:51 crc kubenswrapper[4807]: I1205 12:33:51.325216 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.363836468 podStartE2EDuration="5.325178645s" podCreationTimestamp="2025-12-05 12:33:46 +0000 UTC" firstStartedPulling="2025-12-05 12:33:47.157519843 +0000 UTC m=+1656.651383112" lastFinishedPulling="2025-12-05 12:33:50.11886202 +0000 UTC m=+1659.612725289" observedRunningTime="2025-12-05 12:33:51.315762895 +0000 UTC m=+1660.809626174" watchObservedRunningTime="2025-12-05 12:33:51.325178645 +0000 UTC m=+1660.819041914" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.304037 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.832211 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h"] Dec 05 12:33:52 crc kubenswrapper[4807]: E1205 12:33:52.832567 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1819ab6b-5aab-47ee-ac2d-8afad8ec0b36" containerName="mariadb-account-create-update" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.832584 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1819ab6b-5aab-47ee-ac2d-8afad8ec0b36" containerName="mariadb-account-create-update" Dec 05 12:33:52 crc kubenswrapper[4807]: E1205 12:33:52.832599 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb80014-c9a5-44e0-92f4-8f768ed9ed25" containerName="mariadb-database-create" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.832606 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb80014-c9a5-44e0-92f4-8f768ed9ed25" containerName="mariadb-database-create" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.832756 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1819ab6b-5aab-47ee-ac2d-8afad8ec0b36" containerName="mariadb-account-create-update" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.832774 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb80014-c9a5-44e0-92f4-8f768ed9ed25" containerName="mariadb-database-create" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.833301 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.835963 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-qd28g" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.836040 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.840408 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h"] Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.987666 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbh7q\" (UniqueName: \"kubernetes.io/projected/5efeb544-71b0-4344-9c6d-178ea772dcce-kube-api-access-fbh7q\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.988200 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-config-data\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.988299 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:52 crc kubenswrapper[4807]: I1205 12:33:52.988517 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-db-sync-config-data\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.090567 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbh7q\" (UniqueName: \"kubernetes.io/projected/5efeb544-71b0-4344-9c6d-178ea772dcce-kube-api-access-fbh7q\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.090675 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-config-data\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.091569 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.091757 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-db-sync-config-data\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.095993 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-config-data\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.097102 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-db-sync-config-data\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.097343 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.129316 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbh7q\" (UniqueName: \"kubernetes.io/projected/5efeb544-71b0-4344-9c6d-178ea772dcce-kube-api-access-fbh7q\") pod \"watcher-kuttl-db-sync-jsd7h\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.151147 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:53 crc kubenswrapper[4807]: I1205 12:33:53.588178 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h"] Dec 05 12:33:54 crc kubenswrapper[4807]: I1205 12:33:54.344787 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" event={"ID":"5efeb544-71b0-4344-9c6d-178ea772dcce","Type":"ContainerStarted","Data":"79e5a776728cf67d2ffa0d5beb3b612daea3485a2ca450ae59b037e174fec324"} Dec 05 12:33:54 crc kubenswrapper[4807]: I1205 12:33:54.345144 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" event={"ID":"5efeb544-71b0-4344-9c6d-178ea772dcce","Type":"ContainerStarted","Data":"7a5a2ef12d56716190614a676780d7ae9e4744d15197e8a74e3629b2d838510c"} Dec 05 12:33:54 crc kubenswrapper[4807]: I1205 12:33:54.363153 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" podStartSLOduration=2.363134156 podStartE2EDuration="2.363134156s" podCreationTimestamp="2025-12-05 12:33:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:33:54.3587962 +0000 UTC m=+1663.852659469" watchObservedRunningTime="2025-12-05 12:33:54.363134156 +0000 UTC m=+1663.856997425" Dec 05 12:33:56 crc kubenswrapper[4807]: I1205 12:33:56.366414 4807 generic.go:334] "Generic (PLEG): container finished" podID="5efeb544-71b0-4344-9c6d-178ea772dcce" containerID="79e5a776728cf67d2ffa0d5beb3b612daea3485a2ca450ae59b037e174fec324" exitCode=0 Dec 05 12:33:56 crc kubenswrapper[4807]: I1205 12:33:56.366759 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" event={"ID":"5efeb544-71b0-4344-9c6d-178ea772dcce","Type":"ContainerDied","Data":"79e5a776728cf67d2ffa0d5beb3b612daea3485a2ca450ae59b037e174fec324"} Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.783484 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.877747 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbh7q\" (UniqueName: \"kubernetes.io/projected/5efeb544-71b0-4344-9c6d-178ea772dcce-kube-api-access-fbh7q\") pod \"5efeb544-71b0-4344-9c6d-178ea772dcce\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.877855 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-config-data\") pod \"5efeb544-71b0-4344-9c6d-178ea772dcce\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.877906 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-combined-ca-bundle\") pod \"5efeb544-71b0-4344-9c6d-178ea772dcce\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.877961 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-db-sync-config-data\") pod \"5efeb544-71b0-4344-9c6d-178ea772dcce\" (UID: \"5efeb544-71b0-4344-9c6d-178ea772dcce\") " Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.884054 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5efeb544-71b0-4344-9c6d-178ea772dcce" (UID: "5efeb544-71b0-4344-9c6d-178ea772dcce"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.888247 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5efeb544-71b0-4344-9c6d-178ea772dcce-kube-api-access-fbh7q" (OuterVolumeSpecName: "kube-api-access-fbh7q") pod "5efeb544-71b0-4344-9c6d-178ea772dcce" (UID: "5efeb544-71b0-4344-9c6d-178ea772dcce"). InnerVolumeSpecName "kube-api-access-fbh7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.907851 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5efeb544-71b0-4344-9c6d-178ea772dcce" (UID: "5efeb544-71b0-4344-9c6d-178ea772dcce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.926087 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-config-data" (OuterVolumeSpecName: "config-data") pod "5efeb544-71b0-4344-9c6d-178ea772dcce" (UID: "5efeb544-71b0-4344-9c6d-178ea772dcce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.980189 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbh7q\" (UniqueName: \"kubernetes.io/projected/5efeb544-71b0-4344-9c6d-178ea772dcce-kube-api-access-fbh7q\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.980217 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.980227 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:57 crc kubenswrapper[4807]: I1205 12:33:57.980235 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5efeb544-71b0-4344-9c6d-178ea772dcce-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.391080 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" event={"ID":"5efeb544-71b0-4344-9c6d-178ea772dcce","Type":"ContainerDied","Data":"7a5a2ef12d56716190614a676780d7ae9e4744d15197e8a74e3629b2d838510c"} Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.391128 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a5a2ef12d56716190614a676780d7ae9e4744d15197e8a74e3629b2d838510c" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.391140 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.774873 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:33:58 crc kubenswrapper[4807]: E1205 12:33:58.775299 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5efeb544-71b0-4344-9c6d-178ea772dcce" containerName="watcher-kuttl-db-sync" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.775322 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5efeb544-71b0-4344-9c6d-178ea772dcce" containerName="watcher-kuttl-db-sync" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.775544 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5efeb544-71b0-4344-9c6d-178ea772dcce" containerName="watcher-kuttl-db-sync" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.776264 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.782669 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-qd28g" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.782907 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.795515 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.880594 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.883495 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.895842 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.898088 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.904823 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.905128 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.905130 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.905254 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.905343 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.905384 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.905447 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2xs9\" (UniqueName: \"kubernetes.io/projected/5fb20b34-0143-45c5-b978-bdeee50a1332-kube-api-access-t2xs9\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.905549 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb20b34-0143-45c5-b978-bdeee50a1332-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.933691 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.956136 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.984572 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.985931 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:58 crc kubenswrapper[4807]: I1205 12:33:58.994781 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.007617 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.007664 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.007698 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb20b34-0143-45c5-b978-bdeee50a1332-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.008767 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.008905 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x597s\" (UniqueName: \"kubernetes.io/projected/3b467760-34fa-487e-81e5-2cf6ab78288e-kube-api-access-x597s\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.008933 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-logs\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.008966 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b467760-34fa-487e-81e5-2cf6ab78288e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.008988 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009012 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zwlw\" (UniqueName: \"kubernetes.io/projected/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-kube-api-access-5zwlw\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009088 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009456 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009556 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009575 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009664 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009694 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009720 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.009762 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2xs9\" (UniqueName: \"kubernetes.io/projected/5fb20b34-0143-45c5-b978-bdeee50a1332-kube-api-access-t2xs9\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.010810 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb20b34-0143-45c5-b978-bdeee50a1332-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.017471 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.017717 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.018935 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.029842 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.030609 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2xs9\" (UniqueName: \"kubernetes.io/projected/5fb20b34-0143-45c5-b978-bdeee50a1332-kube-api-access-t2xs9\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.096247 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.110927 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.110978 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111027 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111062 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzfmr\" (UniqueName: \"kubernetes.io/projected/8ad6e641-398f-482c-a55c-47b73e620c13-kube-api-access-tzfmr\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111100 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111124 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x597s\" (UniqueName: \"kubernetes.io/projected/3b467760-34fa-487e-81e5-2cf6ab78288e-kube-api-access-x597s\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111155 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-logs\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111180 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111206 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b467760-34fa-487e-81e5-2cf6ab78288e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111229 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111250 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zwlw\" (UniqueName: \"kubernetes.io/projected/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-kube-api-access-5zwlw\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111277 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111314 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad6e641-398f-482c-a55c-47b73e620c13-logs\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111339 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111358 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111424 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.111494 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.112987 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b467760-34fa-487e-81e5-2cf6ab78288e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.114353 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-logs\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.123310 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.124031 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.124618 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.127199 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.127323 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.129914 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.130576 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.143311 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x597s\" (UniqueName: \"kubernetes.io/projected/3b467760-34fa-487e-81e5-2cf6ab78288e-kube-api-access-x597s\") pod \"watcher-kuttl-applier-0\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.145939 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zwlw\" (UniqueName: \"kubernetes.io/projected/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-kube-api-access-5zwlw\") pod \"watcher-kuttl-api-0\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.213801 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.213874 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.213932 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzfmr\" (UniqueName: \"kubernetes.io/projected/8ad6e641-398f-482c-a55c-47b73e620c13-kube-api-access-tzfmr\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.213964 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.213992 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.214041 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad6e641-398f-482c-a55c-47b73e620c13-logs\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.216689 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad6e641-398f-482c-a55c-47b73e620c13-logs\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.230476 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.230476 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.232637 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.233627 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.240575 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzfmr\" (UniqueName: \"kubernetes.io/projected/8ad6e641-398f-482c-a55c-47b73e620c13-kube-api-access-tzfmr\") pod \"watcher-kuttl-api-1\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.261625 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.270922 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.319066 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.597043 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.765230 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:33:59 crc kubenswrapper[4807]: W1205 12:33:59.774823 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c1a2310_52d5_4d8a_a7be_fe64d32f835b.slice/crio-b127499bb3956beb4c7ff3550bface45893c6c64119590a3c2f25a860c334ebb WatchSource:0}: Error finding container b127499bb3956beb4c7ff3550bface45893c6c64119590a3c2f25a860c334ebb: Status 404 returned error can't find the container with id b127499bb3956beb4c7ff3550bface45893c6c64119590a3c2f25a860c334ebb Dec 05 12:33:59 crc kubenswrapper[4807]: W1205 12:33:59.861696 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b467760_34fa_487e_81e5_2cf6ab78288e.slice/crio-bd5901b36936e29226d553bc66791388a3856d2b0abe2616c40b6df72055e8bf WatchSource:0}: Error finding container bd5901b36936e29226d553bc66791388a3856d2b0abe2616c40b6df72055e8bf: Status 404 returned error can't find the container with id bd5901b36936e29226d553bc66791388a3856d2b0abe2616c40b6df72055e8bf Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.862842 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:33:59 crc kubenswrapper[4807]: W1205 12:33:59.881276 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad6e641_398f_482c_a55c_47b73e620c13.slice/crio-538aa35bded866cdeaf1db4b78d14cfe8e01989f85f57982fd3867b88f2a657b WatchSource:0}: Error finding container 538aa35bded866cdeaf1db4b78d14cfe8e01989f85f57982fd3867b88f2a657b: Status 404 returned error can't find the container with id 538aa35bded866cdeaf1db4b78d14cfe8e01989f85f57982fd3867b88f2a657b Dec 05 12:33:59 crc kubenswrapper[4807]: I1205 12:33:59.888665 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.424809 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3b467760-34fa-487e-81e5-2cf6ab78288e","Type":"ContainerStarted","Data":"3abef8be1c6e87462494477b9fe93163753d147da1113879b208b25e53fd2a45"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.424883 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3b467760-34fa-487e-81e5-2cf6ab78288e","Type":"ContainerStarted","Data":"bd5901b36936e29226d553bc66791388a3856d2b0abe2616c40b6df72055e8bf"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.427120 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"1c1a2310-52d5-4d8a-a7be-fe64d32f835b","Type":"ContainerStarted","Data":"5606fdc00991212b095f921c4166248f8dee2dd74b358d83f4c1ba6a34f543f4"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.427170 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"1c1a2310-52d5-4d8a-a7be-fe64d32f835b","Type":"ContainerStarted","Data":"b127499bb3956beb4c7ff3550bface45893c6c64119590a3c2f25a860c334ebb"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.428886 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"8ad6e641-398f-482c-a55c-47b73e620c13","Type":"ContainerStarted","Data":"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.428921 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"8ad6e641-398f-482c-a55c-47b73e620c13","Type":"ContainerStarted","Data":"538aa35bded866cdeaf1db4b78d14cfe8e01989f85f57982fd3867b88f2a657b"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.430583 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"5fb20b34-0143-45c5-b978-bdeee50a1332","Type":"ContainerStarted","Data":"39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.430607 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"5fb20b34-0143-45c5-b978-bdeee50a1332","Type":"ContainerStarted","Data":"c2ad89a67c962348b5c6ec8ae05dea35e7b918a8197eff59379a68455de8f07a"} Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.456149 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.456130834 podStartE2EDuration="2.456130834s" podCreationTimestamp="2025-12-05 12:33:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:00.441570129 +0000 UTC m=+1669.935433408" watchObservedRunningTime="2025-12-05 12:34:00.456130834 +0000 UTC m=+1669.949994103" Dec 05 12:34:00 crc kubenswrapper[4807]: I1205 12:34:00.468558 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.468541216 podStartE2EDuration="2.468541216s" podCreationTimestamp="2025-12-05 12:33:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:00.457591609 +0000 UTC m=+1669.951454878" watchObservedRunningTime="2025-12-05 12:34:00.468541216 +0000 UTC m=+1669.962404485" Dec 05 12:34:01 crc kubenswrapper[4807]: I1205 12:34:01.242736 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:34:01 crc kubenswrapper[4807]: E1205 12:34:01.243098 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:34:01 crc kubenswrapper[4807]: I1205 12:34:01.440582 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"1c1a2310-52d5-4d8a-a7be-fe64d32f835b","Type":"ContainerStarted","Data":"6693d76d812cdae664706f3dad6538c7ce4a7a5cf969ad3701dd8ea7876aa189"} Dec 05 12:34:01 crc kubenswrapper[4807]: I1205 12:34:01.440940 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:01 crc kubenswrapper[4807]: I1205 12:34:01.442862 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"8ad6e641-398f-482c-a55c-47b73e620c13","Type":"ContainerStarted","Data":"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff"} Dec 05 12:34:01 crc kubenswrapper[4807]: I1205 12:34:01.468170 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=3.4681511609999998 podStartE2EDuration="3.468151161s" podCreationTimestamp="2025-12-05 12:33:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:01.464198936 +0000 UTC m=+1670.958062215" watchObservedRunningTime="2025-12-05 12:34:01.468151161 +0000 UTC m=+1670.962014430" Dec 05 12:34:01 crc kubenswrapper[4807]: I1205 12:34:01.490161 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-1" podStartSLOduration=3.490142016 podStartE2EDuration="3.490142016s" podCreationTimestamp="2025-12-05 12:33:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:01.480560363 +0000 UTC m=+1670.974423632" watchObservedRunningTime="2025-12-05 12:34:01.490142016 +0000 UTC m=+1670.984005285" Dec 05 12:34:02 crc kubenswrapper[4807]: I1205 12:34:02.450809 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:03 crc kubenswrapper[4807]: I1205 12:34:03.706118 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:04 crc kubenswrapper[4807]: I1205 12:34:04.262112 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:04 crc kubenswrapper[4807]: I1205 12:34:04.272265 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:04 crc kubenswrapper[4807]: I1205 12:34:04.321649 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:04 crc kubenswrapper[4807]: I1205 12:34:04.479330 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:34:04 crc kubenswrapper[4807]: I1205 12:34:04.993220 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.097261 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.123867 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.267066 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.271799 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.298347 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.298702 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.322214 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.436321 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.518919 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.526513 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.529234 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.546297 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:09 crc kubenswrapper[4807]: I1205 12:34:09.549967 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:11 crc kubenswrapper[4807]: I1205 12:34:11.860945 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:11 crc kubenswrapper[4807]: I1205 12:34:11.862740 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-central-agent" containerID="cri-o://5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f" gracePeriod=30 Dec 05 12:34:11 crc kubenswrapper[4807]: I1205 12:34:11.862873 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="proxy-httpd" containerID="cri-o://d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6" gracePeriod=30 Dec 05 12:34:11 crc kubenswrapper[4807]: I1205 12:34:11.862934 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="sg-core" containerID="cri-o://067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6" gracePeriod=30 Dec 05 12:34:11 crc kubenswrapper[4807]: I1205 12:34:11.862964 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-notification-agent" containerID="cri-o://236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669" gracePeriod=30 Dec 05 12:34:11 crc kubenswrapper[4807]: I1205 12:34:11.871980 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.171:3000/\": EOF" Dec 05 12:34:12 crc kubenswrapper[4807]: I1205 12:34:12.569999 4807 generic.go:334] "Generic (PLEG): container finished" podID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerID="d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6" exitCode=0 Dec 05 12:34:12 crc kubenswrapper[4807]: I1205 12:34:12.570042 4807 generic.go:334] "Generic (PLEG): container finished" podID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerID="067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6" exitCode=2 Dec 05 12:34:12 crc kubenswrapper[4807]: I1205 12:34:12.570053 4807 generic.go:334] "Generic (PLEG): container finished" podID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerID="5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f" exitCode=0 Dec 05 12:34:12 crc kubenswrapper[4807]: I1205 12:34:12.570078 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerDied","Data":"d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6"} Dec 05 12:34:12 crc kubenswrapper[4807]: I1205 12:34:12.570107 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerDied","Data":"067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6"} Dec 05 12:34:12 crc kubenswrapper[4807]: I1205 12:34:12.570118 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerDied","Data":"5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f"} Dec 05 12:34:13 crc kubenswrapper[4807]: I1205 12:34:13.235238 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:34:13 crc kubenswrapper[4807]: E1205 12:34:13.235749 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.362604 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.519591 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-log-httpd\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.519703 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-sg-core-conf-yaml\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.520060 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.520860 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-config-data\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.521176 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-scripts\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.521250 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh9kr\" (UniqueName: \"kubernetes.io/projected/ddf023d3-0bd1-4a7d-a622-3442787a0309-kube-api-access-qh9kr\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.521651 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-combined-ca-bundle\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.521705 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-run-httpd\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.521747 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-ceilometer-tls-certs\") pod \"ddf023d3-0bd1-4a7d-a622-3442787a0309\" (UID: \"ddf023d3-0bd1-4a7d-a622-3442787a0309\") " Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.521958 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.522191 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.522219 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ddf023d3-0bd1-4a7d-a622-3442787a0309-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.526264 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf023d3-0bd1-4a7d-a622-3442787a0309-kube-api-access-qh9kr" (OuterVolumeSpecName: "kube-api-access-qh9kr") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "kube-api-access-qh9kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.528664 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-scripts" (OuterVolumeSpecName: "scripts") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.565755 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.574069 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.610574 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.617723 4807 generic.go:334] "Generic (PLEG): container finished" podID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerID="236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669" exitCode=0 Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.617806 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerDied","Data":"236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669"} Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.617844 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ddf023d3-0bd1-4a7d-a622-3442787a0309","Type":"ContainerDied","Data":"5b3f5debe10a56df48378cf8df77d0a0ba580c9876fce8f7b103b79428cd3f7c"} Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.617865 4807 scope.go:117] "RemoveContainer" containerID="d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.618121 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.624806 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.624844 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.624861 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh9kr\" (UniqueName: \"kubernetes.io/projected/ddf023d3-0bd1-4a7d-a622-3442787a0309-kube-api-access-qh9kr\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.624876 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.624888 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.626127 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-config-data" (OuterVolumeSpecName: "config-data") pod "ddf023d3-0bd1-4a7d-a622-3442787a0309" (UID: "ddf023d3-0bd1-4a7d-a622-3442787a0309"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.651262 4807 scope.go:117] "RemoveContainer" containerID="067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.672898 4807 scope.go:117] "RemoveContainer" containerID="236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.696981 4807 scope.go:117] "RemoveContainer" containerID="5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.725875 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf023d3-0bd1-4a7d-a622-3442787a0309-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.745800 4807 scope.go:117] "RemoveContainer" containerID="d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6" Dec 05 12:34:16 crc kubenswrapper[4807]: E1205 12:34:16.747076 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6\": container with ID starting with d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6 not found: ID does not exist" containerID="d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.747124 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6"} err="failed to get container status \"d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6\": rpc error: code = NotFound desc = could not find container \"d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6\": container with ID starting with d341699c4df9b810437222e7224e34bddad8e418de7af1328cbe474f0ee555f6 not found: ID does not exist" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.747162 4807 scope.go:117] "RemoveContainer" containerID="067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6" Dec 05 12:34:16 crc kubenswrapper[4807]: E1205 12:34:16.749199 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6\": container with ID starting with 067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6 not found: ID does not exist" containerID="067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.749240 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6"} err="failed to get container status \"067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6\": rpc error: code = NotFound desc = could not find container \"067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6\": container with ID starting with 067488ca0df7359e0537cf1a5bf1094e6eba85cab331afe5a9e97ae65ac901f6 not found: ID does not exist" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.749255 4807 scope.go:117] "RemoveContainer" containerID="236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669" Dec 05 12:34:16 crc kubenswrapper[4807]: E1205 12:34:16.749632 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669\": container with ID starting with 236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669 not found: ID does not exist" containerID="236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.749677 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669"} err="failed to get container status \"236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669\": rpc error: code = NotFound desc = could not find container \"236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669\": container with ID starting with 236f0c76d491854bff315d430f312e7434359e6856e9e190f6d653a91aadb669 not found: ID does not exist" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.749708 4807 scope.go:117] "RemoveContainer" containerID="5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f" Dec 05 12:34:16 crc kubenswrapper[4807]: E1205 12:34:16.750371 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f\": container with ID starting with 5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f not found: ID does not exist" containerID="5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.750404 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f"} err="failed to get container status \"5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f\": rpc error: code = NotFound desc = could not find container \"5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f\": container with ID starting with 5206e61f38c9a11f3af4e813718682a926f4dbbd8cc304ab36513895de6f522f not found: ID does not exist" Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.952600 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:16 crc kubenswrapper[4807]: I1205 12:34:16.970853 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009089 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:17 crc kubenswrapper[4807]: E1205 12:34:17.009475 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="proxy-httpd" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009495 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="proxy-httpd" Dec 05 12:34:17 crc kubenswrapper[4807]: E1205 12:34:17.009537 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="sg-core" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009547 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="sg-core" Dec 05 12:34:17 crc kubenswrapper[4807]: E1205 12:34:17.009569 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-central-agent" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009577 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-central-agent" Dec 05 12:34:17 crc kubenswrapper[4807]: E1205 12:34:17.009596 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-notification-agent" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009604 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-notification-agent" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009778 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-notification-agent" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009797 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="ceilometer-central-agent" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009816 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="proxy-httpd" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.009833 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" containerName="sg-core" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.011617 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.017004 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.017081 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.017380 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.032278 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.132721 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-config-data\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.132792 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.132869 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.132907 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-log-httpd\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.132950 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snjc2\" (UniqueName: \"kubernetes.io/projected/0c241279-5913-46d2-b9e9-eb8e2151041c-kube-api-access-snjc2\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.132989 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-scripts\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.133012 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-run-httpd\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.133088 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.234794 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-log-httpd\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.234853 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snjc2\" (UniqueName: \"kubernetes.io/projected/0c241279-5913-46d2-b9e9-eb8e2151041c-kube-api-access-snjc2\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.234892 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-scripts\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.234917 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-run-httpd\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.234987 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.235036 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-config-data\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.235067 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.235614 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-log-httpd\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.240223 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.240344 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.240403 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-run-httpd\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.240458 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.242484 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-scripts\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.250255 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.250653 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-config-data\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.252101 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf023d3-0bd1-4a7d-a622-3442787a0309" path="/var/lib/kubelet/pods/ddf023d3-0bd1-4a7d-a622-3442787a0309/volumes" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.254269 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snjc2\" (UniqueName: \"kubernetes.io/projected/0c241279-5913-46d2-b9e9-eb8e2151041c-kube-api-access-snjc2\") pod \"ceilometer-0\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.376304 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.570842 4807 scope.go:117] "RemoveContainer" containerID="ee15e917dcf67b32d377fa50d07108e9e96742e9f36fe8a806958ddd2a54339d" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.617967 4807 scope.go:117] "RemoveContainer" containerID="041deaf4a5ec1a298e141f1ecac69a5ae37c79dd2afb79974462ca98bcc879ca" Dec 05 12:34:17 crc kubenswrapper[4807]: I1205 12:34:17.836860 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:18 crc kubenswrapper[4807]: I1205 12:34:18.646069 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerStarted","Data":"8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555"} Dec 05 12:34:18 crc kubenswrapper[4807]: I1205 12:34:18.646650 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerStarted","Data":"f59648eaed9fc89da3b3268f3a417994b50453d5ee8660903e4f6779430fb0a9"} Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.100421 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.101986 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.118641 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.176470 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-combined-ca-bundle\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.176571 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-logs\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.176611 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-config-data\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.176712 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwz2g\" (UniqueName: \"kubernetes.io/projected/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-kube-api-access-bwz2g\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.176756 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-cert-memcached-mtls\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.176779 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-custom-prometheus-ca\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.278272 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-custom-prometheus-ca\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.278361 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-combined-ca-bundle\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.278431 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-logs\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.278454 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-config-data\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.278515 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwz2g\" (UniqueName: \"kubernetes.io/projected/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-kube-api-access-bwz2g\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.278589 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-cert-memcached-mtls\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.282207 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-logs\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.282997 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-cert-memcached-mtls\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.285316 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-custom-prometheus-ca\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.288044 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-config-data\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.288430 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-combined-ca-bundle\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.302977 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwz2g\" (UniqueName: \"kubernetes.io/projected/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-kube-api-access-bwz2g\") pod \"watcher-kuttl-api-2\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.557185 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:19 crc kubenswrapper[4807]: I1205 12:34:19.662801 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerStarted","Data":"7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796"} Dec 05 12:34:20 crc kubenswrapper[4807]: I1205 12:34:20.003305 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Dec 05 12:34:20 crc kubenswrapper[4807]: W1205 12:34:20.005705 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bda2f86_b502_44b0_94cf_f3d25ddbe5e0.slice/crio-2ce000be99b69f002641544fb4c494cf458e6704d1ee4271818ddb99f836a16b WatchSource:0}: Error finding container 2ce000be99b69f002641544fb4c494cf458e6704d1ee4271818ddb99f836a16b: Status 404 returned error can't find the container with id 2ce000be99b69f002641544fb4c494cf458e6704d1ee4271818ddb99f836a16b Dec 05 12:34:20 crc kubenswrapper[4807]: I1205 12:34:20.673102 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerStarted","Data":"2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2"} Dec 05 12:34:20 crc kubenswrapper[4807]: I1205 12:34:20.677013 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0","Type":"ContainerStarted","Data":"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72"} Dec 05 12:34:20 crc kubenswrapper[4807]: I1205 12:34:20.677059 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0","Type":"ContainerStarted","Data":"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6"} Dec 05 12:34:20 crc kubenswrapper[4807]: I1205 12:34:20.677069 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0","Type":"ContainerStarted","Data":"2ce000be99b69f002641544fb4c494cf458e6704d1ee4271818ddb99f836a16b"} Dec 05 12:34:20 crc kubenswrapper[4807]: I1205 12:34:20.677273 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:20 crc kubenswrapper[4807]: I1205 12:34:20.704835 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-2" podStartSLOduration=1.7048102219999999 podStartE2EDuration="1.704810222s" podCreationTimestamp="2025-12-05 12:34:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:20.696648013 +0000 UTC m=+1690.190511282" watchObservedRunningTime="2025-12-05 12:34:20.704810222 +0000 UTC m=+1690.198673491" Dec 05 12:34:21 crc kubenswrapper[4807]: I1205 12:34:21.687342 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerStarted","Data":"e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba"} Dec 05 12:34:21 crc kubenswrapper[4807]: I1205 12:34:21.713571 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.227621062 podStartE2EDuration="5.71354669s" podCreationTimestamp="2025-12-05 12:34:16 +0000 UTC" firstStartedPulling="2025-12-05 12:34:17.85348147 +0000 UTC m=+1687.347344729" lastFinishedPulling="2025-12-05 12:34:21.339407088 +0000 UTC m=+1690.833270357" observedRunningTime="2025-12-05 12:34:21.706513278 +0000 UTC m=+1691.200376567" watchObservedRunningTime="2025-12-05 12:34:21.71354669 +0000 UTC m=+1691.207409969" Dec 05 12:34:22 crc kubenswrapper[4807]: I1205 12:34:22.695787 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:22 crc kubenswrapper[4807]: I1205 12:34:22.915179 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:24 crc kubenswrapper[4807]: I1205 12:34:24.559053 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:28 crc kubenswrapper[4807]: I1205 12:34:28.235665 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:34:28 crc kubenswrapper[4807]: E1205 12:34:28.236208 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:34:29 crc kubenswrapper[4807]: I1205 12:34:29.559199 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:29 crc kubenswrapper[4807]: I1205 12:34:29.565682 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:29 crc kubenswrapper[4807]: I1205 12:34:29.770832 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:30 crc kubenswrapper[4807]: I1205 12:34:30.814779 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Dec 05 12:34:30 crc kubenswrapper[4807]: I1205 12:34:30.827409 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:34:30 crc kubenswrapper[4807]: I1205 12:34:30.827708 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-kuttl-api-log" containerID="cri-o://5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57" gracePeriod=30 Dec 05 12:34:30 crc kubenswrapper[4807]: I1205 12:34:30.827814 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-api" containerID="cri-o://1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff" gracePeriod=30 Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.657617 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.782643 4807 generic.go:334] "Generic (PLEG): container finished" podID="8ad6e641-398f-482c-a55c-47b73e620c13" containerID="1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff" exitCode=0 Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.782940 4807 generic.go:334] "Generic (PLEG): container finished" podID="8ad6e641-398f-482c-a55c-47b73e620c13" containerID="5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57" exitCode=143 Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.782699 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"8ad6e641-398f-482c-a55c-47b73e620c13","Type":"ContainerDied","Data":"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff"} Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.783086 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"8ad6e641-398f-482c-a55c-47b73e620c13","Type":"ContainerDied","Data":"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57"} Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.783117 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"8ad6e641-398f-482c-a55c-47b73e620c13","Type":"ContainerDied","Data":"538aa35bded866cdeaf1db4b78d14cfe8e01989f85f57982fd3867b88f2a657b"} Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.782698 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.783144 4807 scope.go:117] "RemoveContainer" containerID="1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.783409 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-2" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-kuttl-api-log" containerID="cri-o://c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6" gracePeriod=30 Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.783464 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-2" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-api" containerID="cri-o://e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72" gracePeriod=30 Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.804945 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-config-data\") pod \"8ad6e641-398f-482c-a55c-47b73e620c13\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.805030 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-cert-memcached-mtls\") pod \"8ad6e641-398f-482c-a55c-47b73e620c13\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.805220 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad6e641-398f-482c-a55c-47b73e620c13-logs\") pod \"8ad6e641-398f-482c-a55c-47b73e620c13\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.805296 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-custom-prometheus-ca\") pod \"8ad6e641-398f-482c-a55c-47b73e620c13\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.805334 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-combined-ca-bundle\") pod \"8ad6e641-398f-482c-a55c-47b73e620c13\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.805369 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzfmr\" (UniqueName: \"kubernetes.io/projected/8ad6e641-398f-482c-a55c-47b73e620c13-kube-api-access-tzfmr\") pod \"8ad6e641-398f-482c-a55c-47b73e620c13\" (UID: \"8ad6e641-398f-482c-a55c-47b73e620c13\") " Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.807167 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ad6e641-398f-482c-a55c-47b73e620c13-logs" (OuterVolumeSpecName: "logs") pod "8ad6e641-398f-482c-a55c-47b73e620c13" (UID: "8ad6e641-398f-482c-a55c-47b73e620c13"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.807976 4807 scope.go:117] "RemoveContainer" containerID="5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.810637 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad6e641-398f-482c-a55c-47b73e620c13-kube-api-access-tzfmr" (OuterVolumeSpecName: "kube-api-access-tzfmr") pod "8ad6e641-398f-482c-a55c-47b73e620c13" (UID: "8ad6e641-398f-482c-a55c-47b73e620c13"). InnerVolumeSpecName "kube-api-access-tzfmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.835388 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "8ad6e641-398f-482c-a55c-47b73e620c13" (UID: "8ad6e641-398f-482c-a55c-47b73e620c13"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.842405 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ad6e641-398f-482c-a55c-47b73e620c13" (UID: "8ad6e641-398f-482c-a55c-47b73e620c13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.863661 4807 scope.go:117] "RemoveContainer" containerID="1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff" Dec 05 12:34:31 crc kubenswrapper[4807]: E1205 12:34:31.864612 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff\": container with ID starting with 1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff not found: ID does not exist" containerID="1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.864677 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff"} err="failed to get container status \"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff\": rpc error: code = NotFound desc = could not find container \"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff\": container with ID starting with 1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff not found: ID does not exist" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.864721 4807 scope.go:117] "RemoveContainer" containerID="5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57" Dec 05 12:34:31 crc kubenswrapper[4807]: E1205 12:34:31.865211 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57\": container with ID starting with 5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57 not found: ID does not exist" containerID="5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.865248 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57"} err="failed to get container status \"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57\": rpc error: code = NotFound desc = could not find container \"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57\": container with ID starting with 5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57 not found: ID does not exist" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.865273 4807 scope.go:117] "RemoveContainer" containerID="1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.865783 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff"} err="failed to get container status \"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff\": rpc error: code = NotFound desc = could not find container \"1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff\": container with ID starting with 1998d857a636ece69bbc6f4b2a08580ab79ecb2678df3ae7e79e9077b0c33eff not found: ID does not exist" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.865824 4807 scope.go:117] "RemoveContainer" containerID="5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.866174 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57"} err="failed to get container status \"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57\": rpc error: code = NotFound desc = could not find container \"5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57\": container with ID starting with 5596edc730f82708b863da6699d325a3f85ab85e68390ac263a1ee6c6308ab57 not found: ID does not exist" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.875217 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-config-data" (OuterVolumeSpecName: "config-data") pod "8ad6e641-398f-482c-a55c-47b73e620c13" (UID: "8ad6e641-398f-482c-a55c-47b73e620c13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.901748 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "8ad6e641-398f-482c-a55c-47b73e620c13" (UID: "8ad6e641-398f-482c-a55c-47b73e620c13"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.907309 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad6e641-398f-482c-a55c-47b73e620c13-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.907337 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.907346 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.907355 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzfmr\" (UniqueName: \"kubernetes.io/projected/8ad6e641-398f-482c-a55c-47b73e620c13-kube-api-access-tzfmr\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.907363 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:31 crc kubenswrapper[4807]: I1205 12:34:31.907371 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8ad6e641-398f-482c-a55c-47b73e620c13-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.124503 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.132195 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.606397 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.718144 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-combined-ca-bundle\") pod \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.718201 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwz2g\" (UniqueName: \"kubernetes.io/projected/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-kube-api-access-bwz2g\") pod \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.718230 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-logs\") pod \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.718323 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-custom-prometheus-ca\") pod \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.718363 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-config-data\") pod \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.718411 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-cert-memcached-mtls\") pod \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\" (UID: \"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0\") " Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.719217 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-logs" (OuterVolumeSpecName: "logs") pod "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" (UID: "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.725893 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-kube-api-access-bwz2g" (OuterVolumeSpecName: "kube-api-access-bwz2g") pod "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" (UID: "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0"). InnerVolumeSpecName "kube-api-access-bwz2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.743635 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" (UID: "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.745668 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" (UID: "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.766160 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-config-data" (OuterVolumeSpecName: "config-data") pod "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" (UID: "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.793698 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" (UID: "1bda2f86-b502-44b0-94cf-f3d25ddbe5e0"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.798434 4807 generic.go:334] "Generic (PLEG): container finished" podID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerID="e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72" exitCode=0 Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.798477 4807 generic.go:334] "Generic (PLEG): container finished" podID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerID="c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6" exitCode=143 Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.798540 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0","Type":"ContainerDied","Data":"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72"} Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.798575 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0","Type":"ContainerDied","Data":"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6"} Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.798586 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"1bda2f86-b502-44b0-94cf-f3d25ddbe5e0","Type":"ContainerDied","Data":"2ce000be99b69f002641544fb4c494cf458e6704d1ee4271818ddb99f836a16b"} Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.798604 4807 scope.go:117] "RemoveContainer" containerID="e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.798884 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.821970 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.821999 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.822013 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.822025 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.822034 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwz2g\" (UniqueName: \"kubernetes.io/projected/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-kube-api-access-bwz2g\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.822047 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.834017 4807 scope.go:117] "RemoveContainer" containerID="c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.857982 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.858648 4807 scope.go:117] "RemoveContainer" containerID="e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72" Dec 05 12:34:32 crc kubenswrapper[4807]: E1205 12:34:32.858958 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72\": container with ID starting with e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72 not found: ID does not exist" containerID="e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.858983 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72"} err="failed to get container status \"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72\": rpc error: code = NotFound desc = could not find container \"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72\": container with ID starting with e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72 not found: ID does not exist" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.859003 4807 scope.go:117] "RemoveContainer" containerID="c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6" Dec 05 12:34:32 crc kubenswrapper[4807]: E1205 12:34:32.860771 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6\": container with ID starting with c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6 not found: ID does not exist" containerID="c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.860803 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6"} err="failed to get container status \"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6\": rpc error: code = NotFound desc = could not find container \"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6\": container with ID starting with c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6 not found: ID does not exist" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.860819 4807 scope.go:117] "RemoveContainer" containerID="e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.861142 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72"} err="failed to get container status \"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72\": rpc error: code = NotFound desc = could not find container \"e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72\": container with ID starting with e16f7ec412530027e527c63cc03f3642f16891b778bf75fef61bd623eedcbc72 not found: ID does not exist" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.861168 4807 scope.go:117] "RemoveContainer" containerID="c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.861462 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6"} err="failed to get container status \"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6\": rpc error: code = NotFound desc = could not find container \"c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6\": container with ID starting with c42cbb61cbe4f0d23055a039ed238da77e0ca779c52ca328583c66facfb539e6 not found: ID does not exist" Dec 05 12:34:32 crc kubenswrapper[4807]: I1205 12:34:32.871771 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Dec 05 12:34:33 crc kubenswrapper[4807]: I1205 12:34:33.247124 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" path="/var/lib/kubelet/pods/1bda2f86-b502-44b0-94cf-f3d25ddbe5e0/volumes" Dec 05 12:34:33 crc kubenswrapper[4807]: I1205 12:34:33.248025 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" path="/var/lib/kubelet/pods/8ad6e641-398f-482c-a55c-47b73e620c13/volumes" Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.115022 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.115264 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-kuttl-api-log" containerID="cri-o://5606fdc00991212b095f921c4166248f8dee2dd74b358d83f4c1ba6a34f543f4" gracePeriod=30 Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.115355 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-api" containerID="cri-o://6693d76d812cdae664706f3dad6538c7ce4a7a5cf969ad3701dd8ea7876aa189" gracePeriod=30 Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.597974 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.176:9322/\": read tcp 10.217.0.2:60916->10.217.0.176:9322: read: connection reset by peer" Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.598008 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.0.176:9322/\": read tcp 10.217.0.2:60920->10.217.0.176:9322: read: connection reset by peer" Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.822638 4807 generic.go:334] "Generic (PLEG): container finished" podID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerID="6693d76d812cdae664706f3dad6538c7ce4a7a5cf969ad3701dd8ea7876aa189" exitCode=0 Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.822671 4807 generic.go:334] "Generic (PLEG): container finished" podID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerID="5606fdc00991212b095f921c4166248f8dee2dd74b358d83f4c1ba6a34f543f4" exitCode=143 Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.822694 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"1c1a2310-52d5-4d8a-a7be-fe64d32f835b","Type":"ContainerDied","Data":"6693d76d812cdae664706f3dad6538c7ce4a7a5cf969ad3701dd8ea7876aa189"} Dec 05 12:34:34 crc kubenswrapper[4807]: I1205 12:34:34.822731 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"1c1a2310-52d5-4d8a-a7be-fe64d32f835b","Type":"ContainerDied","Data":"5606fdc00991212b095f921c4166248f8dee2dd74b358d83f4c1ba6a34f543f4"} Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.287598 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h"] Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.302868 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-jsd7h"] Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.335978 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher4079-account-delete-84w6t"] Dec 05 12:34:35 crc kubenswrapper[4807]: E1205 12:34:35.336410 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-api" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336430 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-api" Dec 05 12:34:35 crc kubenswrapper[4807]: E1205 12:34:35.336449 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-api" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336456 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-api" Dec 05 12:34:35 crc kubenswrapper[4807]: E1205 12:34:35.336490 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-kuttl-api-log" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336500 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-kuttl-api-log" Dec 05 12:34:35 crc kubenswrapper[4807]: E1205 12:34:35.336511 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-kuttl-api-log" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336518 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-kuttl-api-log" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336749 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-api" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336774 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-api" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336787 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bda2f86-b502-44b0-94cf-f3d25ddbe5e0" containerName="watcher-kuttl-api-log" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.336800 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad6e641-398f-482c-a55c-47b73e620c13" containerName="watcher-kuttl-api-log" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.337479 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.351247 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher4079-account-delete-84w6t"] Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.361598 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.361884 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="3b467760-34fa-487e-81e5-2cf6ab78288e" containerName="watcher-applier" containerID="cri-o://3abef8be1c6e87462494477b9fe93163753d147da1113879b208b25e53fd2a45" gracePeriod=30 Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.444189 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.444643 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="5fb20b34-0143-45c5-b978-bdeee50a1332" containerName="watcher-decision-engine" containerID="cri-o://39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c" gracePeriod=30 Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.464402 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db86a777-45a9-4885-9b5f-ca2321937d82-operator-scripts\") pod \"watcher4079-account-delete-84w6t\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.464490 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx2tt\" (UniqueName: \"kubernetes.io/projected/db86a777-45a9-4885-9b5f-ca2321937d82-kube-api-access-bx2tt\") pod \"watcher4079-account-delete-84w6t\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.530198 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.565536 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db86a777-45a9-4885-9b5f-ca2321937d82-operator-scripts\") pod \"watcher4079-account-delete-84w6t\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.565654 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx2tt\" (UniqueName: \"kubernetes.io/projected/db86a777-45a9-4885-9b5f-ca2321937d82-kube-api-access-bx2tt\") pod \"watcher4079-account-delete-84w6t\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.566543 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db86a777-45a9-4885-9b5f-ca2321937d82-operator-scripts\") pod \"watcher4079-account-delete-84w6t\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.606461 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx2tt\" (UniqueName: \"kubernetes.io/projected/db86a777-45a9-4885-9b5f-ca2321937d82-kube-api-access-bx2tt\") pod \"watcher4079-account-delete-84w6t\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.666495 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-custom-prometheus-ca\") pod \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.666552 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-logs\") pod \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.666614 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zwlw\" (UniqueName: \"kubernetes.io/projected/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-kube-api-access-5zwlw\") pod \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.666632 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-config-data\") pod \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.666656 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-combined-ca-bundle\") pod \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.666677 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-cert-memcached-mtls\") pod \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\" (UID: \"1c1a2310-52d5-4d8a-a7be-fe64d32f835b\") " Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.668109 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-logs" (OuterVolumeSpecName: "logs") pod "1c1a2310-52d5-4d8a-a7be-fe64d32f835b" (UID: "1c1a2310-52d5-4d8a-a7be-fe64d32f835b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.671982 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-kube-api-access-5zwlw" (OuterVolumeSpecName: "kube-api-access-5zwlw") pod "1c1a2310-52d5-4d8a-a7be-fe64d32f835b" (UID: "1c1a2310-52d5-4d8a-a7be-fe64d32f835b"). InnerVolumeSpecName "kube-api-access-5zwlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.676843 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.722750 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c1a2310-52d5-4d8a-a7be-fe64d32f835b" (UID: "1c1a2310-52d5-4d8a-a7be-fe64d32f835b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.744902 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "1c1a2310-52d5-4d8a-a7be-fe64d32f835b" (UID: "1c1a2310-52d5-4d8a-a7be-fe64d32f835b"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.779479 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zwlw\" (UniqueName: \"kubernetes.io/projected/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-kube-api-access-5zwlw\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.779513 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.779540 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.779552 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.797108 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "1c1a2310-52d5-4d8a-a7be-fe64d32f835b" (UID: "1c1a2310-52d5-4d8a-a7be-fe64d32f835b"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.797177 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-config-data" (OuterVolumeSpecName: "config-data") pod "1c1a2310-52d5-4d8a-a7be-fe64d32f835b" (UID: "1c1a2310-52d5-4d8a-a7be-fe64d32f835b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.841703 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"1c1a2310-52d5-4d8a-a7be-fe64d32f835b","Type":"ContainerDied","Data":"b127499bb3956beb4c7ff3550bface45893c6c64119590a3c2f25a860c334ebb"} Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.841759 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.841767 4807 scope.go:117] "RemoveContainer" containerID="6693d76d812cdae664706f3dad6538c7ce4a7a5cf969ad3701dd8ea7876aa189" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.880664 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.880894 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/1c1a2310-52d5-4d8a-a7be-fe64d32f835b-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.892556 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.901131 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:34:35 crc kubenswrapper[4807]: I1205 12:34:35.912381 4807 scope.go:117] "RemoveContainer" containerID="5606fdc00991212b095f921c4166248f8dee2dd74b358d83f4c1ba6a34f543f4" Dec 05 12:34:36 crc kubenswrapper[4807]: I1205 12:34:36.156706 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher4079-account-delete-84w6t"] Dec 05 12:34:36 crc kubenswrapper[4807]: I1205 12:34:36.851566 4807 generic.go:334] "Generic (PLEG): container finished" podID="db86a777-45a9-4885-9b5f-ca2321937d82" containerID="717435056d17930b4e186aadf19c3235bb2adcbe0eea50456a989c8b797dfb0b" exitCode=0 Dec 05 12:34:36 crc kubenswrapper[4807]: I1205 12:34:36.851639 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" event={"ID":"db86a777-45a9-4885-9b5f-ca2321937d82","Type":"ContainerDied","Data":"717435056d17930b4e186aadf19c3235bb2adcbe0eea50456a989c8b797dfb0b"} Dec 05 12:34:36 crc kubenswrapper[4807]: I1205 12:34:36.851883 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" event={"ID":"db86a777-45a9-4885-9b5f-ca2321937d82","Type":"ContainerStarted","Data":"2c879514417cec4d019a6eeca878211f273523ca989c99d025d31cc717f4284e"} Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.245572 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" path="/var/lib/kubelet/pods/1c1a2310-52d5-4d8a-a7be-fe64d32f835b/volumes" Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.248358 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5efeb544-71b0-4344-9c6d-178ea772dcce" path="/var/lib/kubelet/pods/5efeb544-71b0-4344-9c6d-178ea772dcce/volumes" Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.821916 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.822290 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-central-agent" containerID="cri-o://8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555" gracePeriod=30 Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.822973 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="sg-core" containerID="cri-o://2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2" gracePeriod=30 Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.822967 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="proxy-httpd" containerID="cri-o://e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba" gracePeriod=30 Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.823034 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-notification-agent" containerID="cri-o://7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796" gracePeriod=30 Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.856771 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.863786 4807 generic.go:334] "Generic (PLEG): container finished" podID="3b467760-34fa-487e-81e5-2cf6ab78288e" containerID="3abef8be1c6e87462494477b9fe93163753d147da1113879b208b25e53fd2a45" exitCode=0 Dec 05 12:34:37 crc kubenswrapper[4807]: I1205 12:34:37.863873 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3b467760-34fa-487e-81e5-2cf6ab78288e","Type":"ContainerDied","Data":"3abef8be1c6e87462494477b9fe93163753d147da1113879b208b25e53fd2a45"} Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.338638 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.352040 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.422134 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db86a777-45a9-4885-9b5f-ca2321937d82-operator-scripts\") pod \"db86a777-45a9-4885-9b5f-ca2321937d82\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.422192 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-cert-memcached-mtls\") pod \"3b467760-34fa-487e-81e5-2cf6ab78288e\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.422236 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx2tt\" (UniqueName: \"kubernetes.io/projected/db86a777-45a9-4885-9b5f-ca2321937d82-kube-api-access-bx2tt\") pod \"db86a777-45a9-4885-9b5f-ca2321937d82\" (UID: \"db86a777-45a9-4885-9b5f-ca2321937d82\") " Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.422334 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-combined-ca-bundle\") pod \"3b467760-34fa-487e-81e5-2cf6ab78288e\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.422357 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b467760-34fa-487e-81e5-2cf6ab78288e-logs\") pod \"3b467760-34fa-487e-81e5-2cf6ab78288e\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.422453 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-config-data\") pod \"3b467760-34fa-487e-81e5-2cf6ab78288e\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.422469 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x597s\" (UniqueName: \"kubernetes.io/projected/3b467760-34fa-487e-81e5-2cf6ab78288e-kube-api-access-x597s\") pod \"3b467760-34fa-487e-81e5-2cf6ab78288e\" (UID: \"3b467760-34fa-487e-81e5-2cf6ab78288e\") " Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.424035 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db86a777-45a9-4885-9b5f-ca2321937d82-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db86a777-45a9-4885-9b5f-ca2321937d82" (UID: "db86a777-45a9-4885-9b5f-ca2321937d82"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.424313 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b467760-34fa-487e-81e5-2cf6ab78288e-logs" (OuterVolumeSpecName: "logs") pod "3b467760-34fa-487e-81e5-2cf6ab78288e" (UID: "3b467760-34fa-487e-81e5-2cf6ab78288e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.428459 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db86a777-45a9-4885-9b5f-ca2321937d82-kube-api-access-bx2tt" (OuterVolumeSpecName: "kube-api-access-bx2tt") pod "db86a777-45a9-4885-9b5f-ca2321937d82" (UID: "db86a777-45a9-4885-9b5f-ca2321937d82"). InnerVolumeSpecName "kube-api-access-bx2tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.428700 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b467760-34fa-487e-81e5-2cf6ab78288e-kube-api-access-x597s" (OuterVolumeSpecName: "kube-api-access-x597s") pod "3b467760-34fa-487e-81e5-2cf6ab78288e" (UID: "3b467760-34fa-487e-81e5-2cf6ab78288e"). InnerVolumeSpecName "kube-api-access-x597s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.451773 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b467760-34fa-487e-81e5-2cf6ab78288e" (UID: "3b467760-34fa-487e-81e5-2cf6ab78288e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.465728 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-config-data" (OuterVolumeSpecName: "config-data") pod "3b467760-34fa-487e-81e5-2cf6ab78288e" (UID: "3b467760-34fa-487e-81e5-2cf6ab78288e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.486084 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "3b467760-34fa-487e-81e5-2cf6ab78288e" (UID: "3b467760-34fa-487e-81e5-2cf6ab78288e"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.525021 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.525073 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x597s\" (UniqueName: \"kubernetes.io/projected/3b467760-34fa-487e-81e5-2cf6ab78288e-kube-api-access-x597s\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.525088 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db86a777-45a9-4885-9b5f-ca2321937d82-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.525100 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.525111 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx2tt\" (UniqueName: \"kubernetes.io/projected/db86a777-45a9-4885-9b5f-ca2321937d82-kube-api-access-bx2tt\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.525125 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b467760-34fa-487e-81e5-2cf6ab78288e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.525135 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b467760-34fa-487e-81e5-2cf6ab78288e-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.872828 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" event={"ID":"db86a777-45a9-4885-9b5f-ca2321937d82","Type":"ContainerDied","Data":"2c879514417cec4d019a6eeca878211f273523ca989c99d025d31cc717f4284e"} Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.872868 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c879514417cec4d019a6eeca878211f273523ca989c99d025d31cc717f4284e" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.872865 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4079-account-delete-84w6t" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.875464 4807 generic.go:334] "Generic (PLEG): container finished" podID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerID="e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba" exitCode=0 Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.875489 4807 generic.go:334] "Generic (PLEG): container finished" podID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerID="2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2" exitCode=2 Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.875497 4807 generic.go:334] "Generic (PLEG): container finished" podID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerID="8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555" exitCode=0 Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.875532 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerDied","Data":"e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba"} Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.875568 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerDied","Data":"2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2"} Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.875579 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerDied","Data":"8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555"} Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.877025 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"3b467760-34fa-487e-81e5-2cf6ab78288e","Type":"ContainerDied","Data":"bd5901b36936e29226d553bc66791388a3856d2b0abe2616c40b6df72055e8bf"} Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.877058 4807 scope.go:117] "RemoveContainer" containerID="3abef8be1c6e87462494477b9fe93163753d147da1113879b208b25e53fd2a45" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.877169 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.910387 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:34:38 crc kubenswrapper[4807]: I1205 12:34:38.917417 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.246701 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b467760-34fa-487e-81e5-2cf6ab78288e" path="/var/lib/kubelet/pods/3b467760-34fa-487e-81e5-2cf6ab78288e/volumes" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.562150 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.641948 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-config-data\") pod \"5fb20b34-0143-45c5-b978-bdeee50a1332\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.642052 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-combined-ca-bundle\") pod \"5fb20b34-0143-45c5-b978-bdeee50a1332\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.642102 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb20b34-0143-45c5-b978-bdeee50a1332-logs\") pod \"5fb20b34-0143-45c5-b978-bdeee50a1332\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.642125 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-custom-prometheus-ca\") pod \"5fb20b34-0143-45c5-b978-bdeee50a1332\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.642185 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-cert-memcached-mtls\") pod \"5fb20b34-0143-45c5-b978-bdeee50a1332\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.642276 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2xs9\" (UniqueName: \"kubernetes.io/projected/5fb20b34-0143-45c5-b978-bdeee50a1332-kube-api-access-t2xs9\") pod \"5fb20b34-0143-45c5-b978-bdeee50a1332\" (UID: \"5fb20b34-0143-45c5-b978-bdeee50a1332\") " Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.642841 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fb20b34-0143-45c5-b978-bdeee50a1332-logs" (OuterVolumeSpecName: "logs") pod "5fb20b34-0143-45c5-b978-bdeee50a1332" (UID: "5fb20b34-0143-45c5-b978-bdeee50a1332"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.643352 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb20b34-0143-45c5-b978-bdeee50a1332-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.662554 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fb20b34-0143-45c5-b978-bdeee50a1332-kube-api-access-t2xs9" (OuterVolumeSpecName: "kube-api-access-t2xs9") pod "5fb20b34-0143-45c5-b978-bdeee50a1332" (UID: "5fb20b34-0143-45c5-b978-bdeee50a1332"). InnerVolumeSpecName "kube-api-access-t2xs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.682965 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "5fb20b34-0143-45c5-b978-bdeee50a1332" (UID: "5fb20b34-0143-45c5-b978-bdeee50a1332"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.697822 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fb20b34-0143-45c5-b978-bdeee50a1332" (UID: "5fb20b34-0143-45c5-b978-bdeee50a1332"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.727798 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-config-data" (OuterVolumeSpecName: "config-data") pod "5fb20b34-0143-45c5-b978-bdeee50a1332" (UID: "5fb20b34-0143-45c5-b978-bdeee50a1332"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.744321 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.744378 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2xs9\" (UniqueName: \"kubernetes.io/projected/5fb20b34-0143-45c5-b978-bdeee50a1332-kube-api-access-t2xs9\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.744402 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.744410 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.778684 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "5fb20b34-0143-45c5-b978-bdeee50a1332" (UID: "5fb20b34-0143-45c5-b978-bdeee50a1332"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.846017 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/5fb20b34-0143-45c5-b978-bdeee50a1332-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.886699 4807 generic.go:334] "Generic (PLEG): container finished" podID="5fb20b34-0143-45c5-b978-bdeee50a1332" containerID="39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c" exitCode=0 Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.886750 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"5fb20b34-0143-45c5-b978-bdeee50a1332","Type":"ContainerDied","Data":"39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c"} Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.886777 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.886807 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"5fb20b34-0143-45c5-b978-bdeee50a1332","Type":"ContainerDied","Data":"c2ad89a67c962348b5c6ec8ae05dea35e7b918a8197eff59379a68455de8f07a"} Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.886819 4807 scope.go:117] "RemoveContainer" containerID="39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.921699 4807 scope.go:117] "RemoveContainer" containerID="39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.922672 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:34:39 crc kubenswrapper[4807]: E1205 12:34:39.922979 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c\": container with ID starting with 39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c not found: ID does not exist" containerID="39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.923014 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c"} err="failed to get container status \"39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c\": rpc error: code = NotFound desc = could not find container \"39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c\": container with ID starting with 39a20c98583b76f9898e1c47c0c4d5b1c6b67dbfbf96489e4432ceef3f2f4c2c not found: ID does not exist" Dec 05 12:34:39 crc kubenswrapper[4807]: I1205 12:34:39.933823 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:34:40 crc kubenswrapper[4807]: I1205 12:34:40.387015 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lgcpq"] Dec 05 12:34:40 crc kubenswrapper[4807]: I1205 12:34:40.396205 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-4079-account-create-update-fcz88"] Dec 05 12:34:40 crc kubenswrapper[4807]: I1205 12:34:40.404939 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher4079-account-delete-84w6t"] Dec 05 12:34:40 crc kubenswrapper[4807]: I1205 12:34:40.412430 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-lgcpq"] Dec 05 12:34:40 crc kubenswrapper[4807]: I1205 12:34:40.420289 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-4079-account-create-update-fcz88"] Dec 05 12:34:40 crc kubenswrapper[4807]: I1205 12:34:40.428022 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher4079-account-delete-84w6t"] Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.245788 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1819ab6b-5aab-47ee-ac2d-8afad8ec0b36" path="/var/lib/kubelet/pods/1819ab6b-5aab-47ee-ac2d-8afad8ec0b36/volumes" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.246365 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fb20b34-0143-45c5-b978-bdeee50a1332" path="/var/lib/kubelet/pods/5fb20b34-0143-45c5-b978-bdeee50a1332/volumes" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.246837 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb80014-c9a5-44e0-92f4-8f768ed9ed25" path="/var/lib/kubelet/pods/9eb80014-c9a5-44e0-92f4-8f768ed9ed25/volumes" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.247734 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db86a777-45a9-4885-9b5f-ca2321937d82" path="/var/lib/kubelet/pods/db86a777-45a9-4885-9b5f-ca2321937d82/volumes" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.726316 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-n25l8"] Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.726987 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b467760-34fa-487e-81e5-2cf6ab78288e" containerName="watcher-applier" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727009 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b467760-34fa-487e-81e5-2cf6ab78288e" containerName="watcher-applier" Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.727025 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-kuttl-api-log" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727033 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-kuttl-api-log" Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.727043 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb20b34-0143-45c5-b978-bdeee50a1332" containerName="watcher-decision-engine" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727051 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb20b34-0143-45c5-b978-bdeee50a1332" containerName="watcher-decision-engine" Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.727064 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db86a777-45a9-4885-9b5f-ca2321937d82" containerName="mariadb-account-delete" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727072 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="db86a777-45a9-4885-9b5f-ca2321937d82" containerName="mariadb-account-delete" Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.727095 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-api" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727101 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-api" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727254 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fb20b34-0143-45c5-b978-bdeee50a1332" containerName="watcher-decision-engine" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727263 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b467760-34fa-487e-81e5-2cf6ab78288e" containerName="watcher-applier" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727275 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="db86a777-45a9-4885-9b5f-ca2321937d82" containerName="mariadb-account-delete" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727330 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-kuttl-api-log" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727344 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c1a2310-52d5-4d8a-a7be-fe64d32f835b" containerName="watcher-api" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.727984 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.733514 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-n25l8"] Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.778551 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7f4g\" (UniqueName: \"kubernetes.io/projected/904acf87-6567-4146-84d2-38c23643b7d6-kube-api-access-h7f4g\") pod \"watcher-db-create-n25l8\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.778685 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/904acf87-6567-4146-84d2-38c23643b7d6-operator-scripts\") pod \"watcher-db-create-n25l8\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.822483 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.824831 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2"] Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.825203 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="proxy-httpd" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825227 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="proxy-httpd" Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.825242 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="sg-core" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825250 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="sg-core" Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.825264 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-central-agent" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825271 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-central-agent" Dec 05 12:34:41 crc kubenswrapper[4807]: E1205 12:34:41.825284 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-notification-agent" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825291 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-notification-agent" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825496 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-central-agent" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825599 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="ceilometer-notification-agent" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825613 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="proxy-httpd" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.825629 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerName="sg-core" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.828310 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.833902 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.842342 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2"] Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880024 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-run-httpd\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880060 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-scripts\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880123 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-combined-ca-bundle\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880152 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-config-data\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880187 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snjc2\" (UniqueName: \"kubernetes.io/projected/0c241279-5913-46d2-b9e9-eb8e2151041c-kube-api-access-snjc2\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880204 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-sg-core-conf-yaml\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880242 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-ceilometer-tls-certs\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880290 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-log-httpd\") pod \"0c241279-5913-46d2-b9e9-eb8e2151041c\" (UID: \"0c241279-5913-46d2-b9e9-eb8e2151041c\") " Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880570 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddlnk\" (UniqueName: \"kubernetes.io/projected/5a873051-f632-40ff-8086-c1017d8f09c7-kube-api-access-ddlnk\") pod \"watcher-4cf0-account-create-update-6jjf2\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880623 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/904acf87-6567-4146-84d2-38c23643b7d6-operator-scripts\") pod \"watcher-db-create-n25l8\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880693 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7f4g\" (UniqueName: \"kubernetes.io/projected/904acf87-6567-4146-84d2-38c23643b7d6-kube-api-access-h7f4g\") pod \"watcher-db-create-n25l8\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.880755 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a873051-f632-40ff-8086-c1017d8f09c7-operator-scripts\") pod \"watcher-4cf0-account-create-update-6jjf2\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.881620 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.881909 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.888323 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/904acf87-6567-4146-84d2-38c23643b7d6-operator-scripts\") pod \"watcher-db-create-n25l8\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.890926 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c241279-5913-46d2-b9e9-eb8e2151041c-kube-api-access-snjc2" (OuterVolumeSpecName: "kube-api-access-snjc2") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "kube-api-access-snjc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.893662 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-scripts" (OuterVolumeSpecName: "scripts") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.922125 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7f4g\" (UniqueName: \"kubernetes.io/projected/904acf87-6567-4146-84d2-38c23643b7d6-kube-api-access-h7f4g\") pod \"watcher-db-create-n25l8\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.923952 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.943266 4807 generic.go:334] "Generic (PLEG): container finished" podID="0c241279-5913-46d2-b9e9-eb8e2151041c" containerID="7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796" exitCode=0 Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.943315 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerDied","Data":"7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796"} Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.943346 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c241279-5913-46d2-b9e9-eb8e2151041c","Type":"ContainerDied","Data":"f59648eaed9fc89da3b3268f3a417994b50453d5ee8660903e4f6779430fb0a9"} Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.943348 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.943363 4807 scope.go:117] "RemoveContainer" containerID="e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.948910 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.973982 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.979705 4807 scope.go:117] "RemoveContainer" containerID="2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982659 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a873051-f632-40ff-8086-c1017d8f09c7-operator-scripts\") pod \"watcher-4cf0-account-create-update-6jjf2\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982725 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddlnk\" (UniqueName: \"kubernetes.io/projected/5a873051-f632-40ff-8086-c1017d8f09c7-kube-api-access-ddlnk\") pod \"watcher-4cf0-account-create-update-6jjf2\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982845 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982861 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c241279-5913-46d2-b9e9-eb8e2151041c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982869 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982878 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982888 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snjc2\" (UniqueName: \"kubernetes.io/projected/0c241279-5913-46d2-b9e9-eb8e2151041c-kube-api-access-snjc2\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982897 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.982905 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.983985 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a873051-f632-40ff-8086-c1017d8f09c7-operator-scripts\") pod \"watcher-4cf0-account-create-update-6jjf2\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.995136 4807 scope.go:117] "RemoveContainer" containerID="7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.995204 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-config-data" (OuterVolumeSpecName: "config-data") pod "0c241279-5913-46d2-b9e9-eb8e2151041c" (UID: "0c241279-5913-46d2-b9e9-eb8e2151041c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:41 crc kubenswrapper[4807]: I1205 12:34:41.998771 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddlnk\" (UniqueName: \"kubernetes.io/projected/5a873051-f632-40ff-8086-c1017d8f09c7-kube-api-access-ddlnk\") pod \"watcher-4cf0-account-create-update-6jjf2\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.013845 4807 scope.go:117] "RemoveContainer" containerID="8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.029586 4807 scope.go:117] "RemoveContainer" containerID="e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba" Dec 05 12:34:42 crc kubenswrapper[4807]: E1205 12:34:42.030020 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba\": container with ID starting with e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba not found: ID does not exist" containerID="e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.030067 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba"} err="failed to get container status \"e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba\": rpc error: code = NotFound desc = could not find container \"e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba\": container with ID starting with e54be64b45c8a2006c3eca76ad8664af1b22805b6dfd7767642b909b4eed0dba not found: ID does not exist" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.030093 4807 scope.go:117] "RemoveContainer" containerID="2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2" Dec 05 12:34:42 crc kubenswrapper[4807]: E1205 12:34:42.030341 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2\": container with ID starting with 2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2 not found: ID does not exist" containerID="2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.030370 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2"} err="failed to get container status \"2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2\": rpc error: code = NotFound desc = could not find container \"2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2\": container with ID starting with 2fce25a0144b832d97557ffaa8773e31ecb5c74c4475113a591d29e49d65a7a2 not found: ID does not exist" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.030388 4807 scope.go:117] "RemoveContainer" containerID="7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796" Dec 05 12:34:42 crc kubenswrapper[4807]: E1205 12:34:42.030637 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796\": container with ID starting with 7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796 not found: ID does not exist" containerID="7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.030664 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796"} err="failed to get container status \"7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796\": rpc error: code = NotFound desc = could not find container \"7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796\": container with ID starting with 7f020578bd141978ffdcc029fe53a8639a5dc81b5b13a1f5caba59fc62e38796 not found: ID does not exist" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.030681 4807 scope.go:117] "RemoveContainer" containerID="8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555" Dec 05 12:34:42 crc kubenswrapper[4807]: E1205 12:34:42.030877 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555\": container with ID starting with 8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555 not found: ID does not exist" containerID="8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.030904 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555"} err="failed to get container status \"8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555\": rpc error: code = NotFound desc = could not find container \"8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555\": container with ID starting with 8a3413db034b73f89c86d8abbb43735f6079675b3b5e9391d06ca612aa024555 not found: ID does not exist" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.084693 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c241279-5913-46d2-b9e9-eb8e2151041c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.118192 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.160467 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.314635 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.337503 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.357721 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.366622 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.369912 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.370652 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.371628 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.393447 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.493381 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.493726 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-config-data\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.493763 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-run-httpd\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.493814 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j8ck\" (UniqueName: \"kubernetes.io/projected/5cd50e15-557b-4d86-af33-56ce3ace1698-kube-api-access-9j8ck\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.494001 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.494058 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-log-httpd\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.494076 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-scripts\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.494110 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595213 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-log-httpd\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595270 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-scripts\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595298 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595326 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595348 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-config-data\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595380 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-run-httpd\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595433 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j8ck\" (UniqueName: \"kubernetes.io/projected/5cd50e15-557b-4d86-af33-56ce3ace1698-kube-api-access-9j8ck\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595540 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.595917 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-log-httpd\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.596789 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-run-httpd\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.600431 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.602287 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-n25l8"] Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.604888 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.605155 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-config-data\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.607066 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-scripts\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.613305 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.632606 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j8ck\" (UniqueName: \"kubernetes.io/projected/5cd50e15-557b-4d86-af33-56ce3ace1698-kube-api-access-9j8ck\") pod \"ceilometer-0\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.727955 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:42 crc kubenswrapper[4807]: I1205 12:34:42.783551 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2"] Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:42.959716 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" event={"ID":"5a873051-f632-40ff-8086-c1017d8f09c7","Type":"ContainerStarted","Data":"6d362b7ff6b7ec9e8ac23c1e87c8c55384e1ee992478a5b9f870a842032b39df"} Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:42.961934 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-n25l8" event={"ID":"904acf87-6567-4146-84d2-38c23643b7d6","Type":"ContainerStarted","Data":"2c50c3f64b92e30a1cb34b33dabc5d960cb04ceccd85e1ebc40cff7e1cf7bea4"} Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:42.961975 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-n25l8" event={"ID":"904acf87-6567-4146-84d2-38c23643b7d6","Type":"ContainerStarted","Data":"71a6aabfac68bd826093fbd471433fab88f403cae8a7b57c0b6f48fbe4a5a042"} Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:42.982380 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-db-create-n25l8" podStartSLOduration=1.982365684 podStartE2EDuration="1.982365684s" podCreationTimestamp="2025-12-05 12:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:42.981039372 +0000 UTC m=+1712.474902651" watchObservedRunningTime="2025-12-05 12:34:42.982365684 +0000 UTC m=+1712.476228953" Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.236755 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:34:43 crc kubenswrapper[4807]: E1205 12:34:43.236987 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.244967 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c241279-5913-46d2-b9e9-eb8e2151041c" path="/var/lib/kubelet/pods/0c241279-5913-46d2-b9e9-eb8e2151041c/volumes" Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.865003 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:34:43 crc kubenswrapper[4807]: W1205 12:34:43.866755 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cd50e15_557b_4d86_af33_56ce3ace1698.slice/crio-a7a19177af34416b46342ea9dd7c388c576e32ae643988561841577df9ddb562 WatchSource:0}: Error finding container a7a19177af34416b46342ea9dd7c388c576e32ae643988561841577df9ddb562: Status 404 returned error can't find the container with id a7a19177af34416b46342ea9dd7c388c576e32ae643988561841577df9ddb562 Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.973966 4807 generic.go:334] "Generic (PLEG): container finished" podID="5a873051-f632-40ff-8086-c1017d8f09c7" containerID="d7d356063957afa6a96ad91c0b931778ba3ab806a4274ed80b5f482b721a0dfd" exitCode=0 Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.974042 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" event={"ID":"5a873051-f632-40ff-8086-c1017d8f09c7","Type":"ContainerDied","Data":"d7d356063957afa6a96ad91c0b931778ba3ab806a4274ed80b5f482b721a0dfd"} Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.975515 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerStarted","Data":"a7a19177af34416b46342ea9dd7c388c576e32ae643988561841577df9ddb562"} Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.977633 4807 generic.go:334] "Generic (PLEG): container finished" podID="904acf87-6567-4146-84d2-38c23643b7d6" containerID="2c50c3f64b92e30a1cb34b33dabc5d960cb04ceccd85e1ebc40cff7e1cf7bea4" exitCode=0 Dec 05 12:34:43 crc kubenswrapper[4807]: I1205 12:34:43.977663 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-n25l8" event={"ID":"904acf87-6567-4146-84d2-38c23643b7d6","Type":"ContainerDied","Data":"2c50c3f64b92e30a1cb34b33dabc5d960cb04ceccd85e1ebc40cff7e1cf7bea4"} Dec 05 12:34:44 crc kubenswrapper[4807]: I1205 12:34:44.989048 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerStarted","Data":"59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21"} Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.380339 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.387662 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.447292 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/904acf87-6567-4146-84d2-38c23643b7d6-operator-scripts\") pod \"904acf87-6567-4146-84d2-38c23643b7d6\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.447346 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7f4g\" (UniqueName: \"kubernetes.io/projected/904acf87-6567-4146-84d2-38c23643b7d6-kube-api-access-h7f4g\") pod \"904acf87-6567-4146-84d2-38c23643b7d6\" (UID: \"904acf87-6567-4146-84d2-38c23643b7d6\") " Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.447390 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a873051-f632-40ff-8086-c1017d8f09c7-operator-scripts\") pod \"5a873051-f632-40ff-8086-c1017d8f09c7\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.447577 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddlnk\" (UniqueName: \"kubernetes.io/projected/5a873051-f632-40ff-8086-c1017d8f09c7-kube-api-access-ddlnk\") pod \"5a873051-f632-40ff-8086-c1017d8f09c7\" (UID: \"5a873051-f632-40ff-8086-c1017d8f09c7\") " Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.448395 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904acf87-6567-4146-84d2-38c23643b7d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "904acf87-6567-4146-84d2-38c23643b7d6" (UID: "904acf87-6567-4146-84d2-38c23643b7d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.448806 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a873051-f632-40ff-8086-c1017d8f09c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a873051-f632-40ff-8086-c1017d8f09c7" (UID: "5a873051-f632-40ff-8086-c1017d8f09c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.458837 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/904acf87-6567-4146-84d2-38c23643b7d6-kube-api-access-h7f4g" (OuterVolumeSpecName: "kube-api-access-h7f4g") pod "904acf87-6567-4146-84d2-38c23643b7d6" (UID: "904acf87-6567-4146-84d2-38c23643b7d6"). InnerVolumeSpecName "kube-api-access-h7f4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.464763 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a873051-f632-40ff-8086-c1017d8f09c7-kube-api-access-ddlnk" (OuterVolumeSpecName: "kube-api-access-ddlnk") pod "5a873051-f632-40ff-8086-c1017d8f09c7" (UID: "5a873051-f632-40ff-8086-c1017d8f09c7"). InnerVolumeSpecName "kube-api-access-ddlnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.549729 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/904acf87-6567-4146-84d2-38c23643b7d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.549927 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7f4g\" (UniqueName: \"kubernetes.io/projected/904acf87-6567-4146-84d2-38c23643b7d6-kube-api-access-h7f4g\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.549992 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a873051-f632-40ff-8086-c1017d8f09c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:45 crc kubenswrapper[4807]: I1205 12:34:45.550043 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddlnk\" (UniqueName: \"kubernetes.io/projected/5a873051-f632-40ff-8086-c1017d8f09c7-kube-api-access-ddlnk\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.007378 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" event={"ID":"5a873051-f632-40ff-8086-c1017d8f09c7","Type":"ContainerDied","Data":"6d362b7ff6b7ec9e8ac23c1e87c8c55384e1ee992478a5b9f870a842032b39df"} Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.007665 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d362b7ff6b7ec9e8ac23c1e87c8c55384e1ee992478a5b9f870a842032b39df" Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.007693 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2" Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.023945 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerStarted","Data":"c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4"} Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.023993 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerStarted","Data":"3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4"} Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.025748 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-n25l8" event={"ID":"904acf87-6567-4146-84d2-38c23643b7d6","Type":"ContainerDied","Data":"71a6aabfac68bd826093fbd471433fab88f403cae8a7b57c0b6f48fbe4a5a042"} Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.025779 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71a6aabfac68bd826093fbd471433fab88f403cae8a7b57c0b6f48fbe4a5a042" Dec 05 12:34:46 crc kubenswrapper[4807]: I1205 12:34:46.025895 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-n25l8" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.224409 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw"] Dec 05 12:34:47 crc kubenswrapper[4807]: E1205 12:34:47.225024 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="904acf87-6567-4146-84d2-38c23643b7d6" containerName="mariadb-database-create" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.225036 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="904acf87-6567-4146-84d2-38c23643b7d6" containerName="mariadb-database-create" Dec 05 12:34:47 crc kubenswrapper[4807]: E1205 12:34:47.225050 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a873051-f632-40ff-8086-c1017d8f09c7" containerName="mariadb-account-create-update" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.225056 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a873051-f632-40ff-8086-c1017d8f09c7" containerName="mariadb-account-create-update" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.225223 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a873051-f632-40ff-8086-c1017d8f09c7" containerName="mariadb-account-create-update" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.225235 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="904acf87-6567-4146-84d2-38c23643b7d6" containerName="mariadb-database-create" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.225890 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.231602 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.231641 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-xvs4w" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.246491 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw"] Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.279683 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m5bd\" (UniqueName: \"kubernetes.io/projected/73daef9b-354f-4347-bbcd-32b8375a579e-kube-api-access-8m5bd\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.279744 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.279831 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-config-data\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.279881 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-db-sync-config-data\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.380902 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.381014 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-config-data\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.381083 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-db-sync-config-data\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.381129 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m5bd\" (UniqueName: \"kubernetes.io/projected/73daef9b-354f-4347-bbcd-32b8375a579e-kube-api-access-8m5bd\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.389251 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-db-sync-config-data\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.389609 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-config-data\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.396252 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.407993 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m5bd\" (UniqueName: \"kubernetes.io/projected/73daef9b-354f-4347-bbcd-32b8375a579e-kube-api-access-8m5bd\") pod \"watcher-kuttl-db-sync-nsvdw\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:47 crc kubenswrapper[4807]: I1205 12:34:47.543831 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:48 crc kubenswrapper[4807]: I1205 12:34:48.017938 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw"] Dec 05 12:34:48 crc kubenswrapper[4807]: W1205 12:34:48.025682 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73daef9b_354f_4347_bbcd_32b8375a579e.slice/crio-c7a0a0e16a5ee26ab130b9b72c9a171ad9669b098a76b94beedf7ab3a262a440 WatchSource:0}: Error finding container c7a0a0e16a5ee26ab130b9b72c9a171ad9669b098a76b94beedf7ab3a262a440: Status 404 returned error can't find the container with id c7a0a0e16a5ee26ab130b9b72c9a171ad9669b098a76b94beedf7ab3a262a440 Dec 05 12:34:48 crc kubenswrapper[4807]: I1205 12:34:48.042706 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" event={"ID":"73daef9b-354f-4347-bbcd-32b8375a579e","Type":"ContainerStarted","Data":"c7a0a0e16a5ee26ab130b9b72c9a171ad9669b098a76b94beedf7ab3a262a440"} Dec 05 12:34:48 crc kubenswrapper[4807]: I1205 12:34:48.046341 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerStarted","Data":"b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8"} Dec 05 12:34:48 crc kubenswrapper[4807]: I1205 12:34:48.046663 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:34:48 crc kubenswrapper[4807]: I1205 12:34:48.083958 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=3.099146597 podStartE2EDuration="6.083937774s" podCreationTimestamp="2025-12-05 12:34:42 +0000 UTC" firstStartedPulling="2025-12-05 12:34:43.868888749 +0000 UTC m=+1713.362752018" lastFinishedPulling="2025-12-05 12:34:46.853679926 +0000 UTC m=+1716.347543195" observedRunningTime="2025-12-05 12:34:48.072788592 +0000 UTC m=+1717.566651861" watchObservedRunningTime="2025-12-05 12:34:48.083937774 +0000 UTC m=+1717.577801043" Dec 05 12:34:49 crc kubenswrapper[4807]: I1205 12:34:49.056120 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" event={"ID":"73daef9b-354f-4347-bbcd-32b8375a579e","Type":"ContainerStarted","Data":"c38e3e7a05d0f6dc7d2e811b8fd0f33138de0bba3977593802d46d8d1e58971e"} Dec 05 12:34:49 crc kubenswrapper[4807]: I1205 12:34:49.081208 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" podStartSLOduration=2.081186553 podStartE2EDuration="2.081186553s" podCreationTimestamp="2025-12-05 12:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:49.075382621 +0000 UTC m=+1718.569245880" watchObservedRunningTime="2025-12-05 12:34:49.081186553 +0000 UTC m=+1718.575049822" Dec 05 12:34:51 crc kubenswrapper[4807]: I1205 12:34:51.073926 4807 generic.go:334] "Generic (PLEG): container finished" podID="73daef9b-354f-4347-bbcd-32b8375a579e" containerID="c38e3e7a05d0f6dc7d2e811b8fd0f33138de0bba3977593802d46d8d1e58971e" exitCode=0 Dec 05 12:34:51 crc kubenswrapper[4807]: I1205 12:34:51.073993 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" event={"ID":"73daef9b-354f-4347-bbcd-32b8375a579e","Type":"ContainerDied","Data":"c38e3e7a05d0f6dc7d2e811b8fd0f33138de0bba3977593802d46d8d1e58971e"} Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.441186 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.575658 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-combined-ca-bundle\") pod \"73daef9b-354f-4347-bbcd-32b8375a579e\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.575761 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-db-sync-config-data\") pod \"73daef9b-354f-4347-bbcd-32b8375a579e\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.575839 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m5bd\" (UniqueName: \"kubernetes.io/projected/73daef9b-354f-4347-bbcd-32b8375a579e-kube-api-access-8m5bd\") pod \"73daef9b-354f-4347-bbcd-32b8375a579e\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.575899 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-config-data\") pod \"73daef9b-354f-4347-bbcd-32b8375a579e\" (UID: \"73daef9b-354f-4347-bbcd-32b8375a579e\") " Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.585736 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73daef9b-354f-4347-bbcd-32b8375a579e-kube-api-access-8m5bd" (OuterVolumeSpecName: "kube-api-access-8m5bd") pod "73daef9b-354f-4347-bbcd-32b8375a579e" (UID: "73daef9b-354f-4347-bbcd-32b8375a579e"). InnerVolumeSpecName "kube-api-access-8m5bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.591773 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "73daef9b-354f-4347-bbcd-32b8375a579e" (UID: "73daef9b-354f-4347-bbcd-32b8375a579e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.620749 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73daef9b-354f-4347-bbcd-32b8375a579e" (UID: "73daef9b-354f-4347-bbcd-32b8375a579e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.635771 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-config-data" (OuterVolumeSpecName: "config-data") pod "73daef9b-354f-4347-bbcd-32b8375a579e" (UID: "73daef9b-354f-4347-bbcd-32b8375a579e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.677284 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.677329 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.677340 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/73daef9b-354f-4347-bbcd-32b8375a579e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:52 crc kubenswrapper[4807]: I1205 12:34:52.677349 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m5bd\" (UniqueName: \"kubernetes.io/projected/73daef9b-354f-4347-bbcd-32b8375a579e-kube-api-access-8m5bd\") on node \"crc\" DevicePath \"\"" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.090261 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" event={"ID":"73daef9b-354f-4347-bbcd-32b8375a579e","Type":"ContainerDied","Data":"c7a0a0e16a5ee26ab130b9b72c9a171ad9669b098a76b94beedf7ab3a262a440"} Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.090318 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a0a0e16a5ee26ab130b9b72c9a171ad9669b098a76b94beedf7ab3a262a440" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.090363 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.357429 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:34:53 crc kubenswrapper[4807]: E1205 12:34:53.357855 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73daef9b-354f-4347-bbcd-32b8375a579e" containerName="watcher-kuttl-db-sync" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.357879 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="73daef9b-354f-4347-bbcd-32b8375a579e" containerName="watcher-kuttl-db-sync" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.358086 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="73daef9b-354f-4347-bbcd-32b8375a579e" containerName="watcher-kuttl-db-sync" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.359175 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.366912 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-xvs4w" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.367305 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.464615 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.478655 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.479733 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.486362 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.491893 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.491966 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.492012 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.492062 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4sb7\" (UniqueName: \"kubernetes.io/projected/410711c8-f736-4f77-ae89-e43fe721e7cd-kube-api-access-v4sb7\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.492089 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410711c8-f736-4f77-ae89-e43fe721e7cd-logs\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.492111 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.504827 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.566591 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.567837 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.570890 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.578770 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595309 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595371 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595399 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv2rj\" (UniqueName: \"kubernetes.io/projected/6735e09d-799c-4aed-ac7a-02e301da235f-kube-api-access-pv2rj\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595418 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595441 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595479 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4sb7\" (UniqueName: \"kubernetes.io/projected/410711c8-f736-4f77-ae89-e43fe721e7cd-kube-api-access-v4sb7\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595500 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410711c8-f736-4f77-ae89-e43fe721e7cd-logs\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595518 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595555 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595574 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.595601 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6735e09d-799c-4aed-ac7a-02e301da235f-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.596870 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410711c8-f736-4f77-ae89-e43fe721e7cd-logs\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.623846 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.624305 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.625165 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.625254 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4sb7\" (UniqueName: \"kubernetes.io/projected/410711c8-f736-4f77-ae89-e43fe721e7cd-kube-api-access-v4sb7\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.629067 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.688913 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697522 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122f1236-de9b-447f-b378-7fa22891f74e-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697583 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv2rj\" (UniqueName: \"kubernetes.io/projected/6735e09d-799c-4aed-ac7a-02e301da235f-kube-api-access-pv2rj\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697607 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697660 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697703 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697730 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697766 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697788 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89t2q\" (UniqueName: \"kubernetes.io/projected/122f1236-de9b-447f-b378-7fa22891f74e-kube-api-access-89t2q\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697814 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6735e09d-799c-4aed-ac7a-02e301da235f-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697842 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.697859 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.698543 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6735e09d-799c-4aed-ac7a-02e301da235f-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.701244 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.704659 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.704758 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.723771 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv2rj\" (UniqueName: \"kubernetes.io/projected/6735e09d-799c-4aed-ac7a-02e301da235f-kube-api-access-pv2rj\") pod \"watcher-kuttl-applier-0\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.814690 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.815682 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.815720 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89t2q\" (UniqueName: \"kubernetes.io/projected/122f1236-de9b-447f-b378-7fa22891f74e-kube-api-access-89t2q\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.815780 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.815809 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.817483 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122f1236-de9b-447f-b378-7fa22891f74e-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.818004 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122f1236-de9b-447f-b378-7fa22891f74e-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.821166 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.821937 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.822662 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.826743 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.827215 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.841038 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89t2q\" (UniqueName: \"kubernetes.io/projected/122f1236-de9b-447f-b378-7fa22891f74e-kube-api-access-89t2q\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:53 crc kubenswrapper[4807]: I1205 12:34:53.892652 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:34:54 crc kubenswrapper[4807]: I1205 12:34:54.228623 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:34:54 crc kubenswrapper[4807]: I1205 12:34:54.362392 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:34:54 crc kubenswrapper[4807]: I1205 12:34:54.470449 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:34:54 crc kubenswrapper[4807]: W1205 12:34:54.497474 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod122f1236_de9b_447f_b378_7fa22891f74e.slice/crio-a047df03065aa05f203c0e71eb6f097e3b6a6d1ff233840ca8e3a7eef8c94493 WatchSource:0}: Error finding container a047df03065aa05f203c0e71eb6f097e3b6a6d1ff233840ca8e3a7eef8c94493: Status 404 returned error can't find the container with id a047df03065aa05f203c0e71eb6f097e3b6a6d1ff233840ca8e3a7eef8c94493 Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.134135 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"122f1236-de9b-447f-b378-7fa22891f74e","Type":"ContainerStarted","Data":"92e613d2405d5672afb26257ed113131f48d3ae321d90f120bd7d380646134c1"} Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.134189 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"122f1236-de9b-447f-b378-7fa22891f74e","Type":"ContainerStarted","Data":"a047df03065aa05f203c0e71eb6f097e3b6a6d1ff233840ca8e3a7eef8c94493"} Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.136675 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6735e09d-799c-4aed-ac7a-02e301da235f","Type":"ContainerStarted","Data":"da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32"} Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.137745 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6735e09d-799c-4aed-ac7a-02e301da235f","Type":"ContainerStarted","Data":"6287e56bc5e87f516b1f41e40b4a59835a721d42e5f7617cb9662f399c5c9cee"} Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.140248 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"410711c8-f736-4f77-ae89-e43fe721e7cd","Type":"ContainerStarted","Data":"79dac9f110987b8941c35b7544a549ef83c6afe470d621093d7551ff45659185"} Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.140388 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"410711c8-f736-4f77-ae89-e43fe721e7cd","Type":"ContainerStarted","Data":"20a268d732dca89e82a6f4b36161c1aac4100e84574aaef120d01f5e3b2139dc"} Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.140477 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"410711c8-f736-4f77-ae89-e43fe721e7cd","Type":"ContainerStarted","Data":"7e189443e35a0db158d2568a741481f5be9af51df14e22a063801db86e42a47e"} Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.143754 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.168929 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.168908672 podStartE2EDuration="2.168908672s" podCreationTimestamp="2025-12-05 12:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:55.164552716 +0000 UTC m=+1724.658415995" watchObservedRunningTime="2025-12-05 12:34:55.168908672 +0000 UTC m=+1724.662771941" Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.189155 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.189139805 podStartE2EDuration="2.189139805s" podCreationTimestamp="2025-12-05 12:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:55.187209657 +0000 UTC m=+1724.681072926" watchObservedRunningTime="2025-12-05 12:34:55.189139805 +0000 UTC m=+1724.683003074" Dec 05 12:34:55 crc kubenswrapper[4807]: I1205 12:34:55.954024 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:34:57 crc kubenswrapper[4807]: I1205 12:34:57.154489 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:34:57 crc kubenswrapper[4807]: I1205 12:34:57.202469 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:34:57 crc kubenswrapper[4807]: I1205 12:34:57.538117 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:57 crc kubenswrapper[4807]: I1205 12:34:57.565619 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=4.565603484 podStartE2EDuration="4.565603484s" podCreationTimestamp="2025-12-05 12:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:34:55.218019467 +0000 UTC m=+1724.711882746" watchObservedRunningTime="2025-12-05 12:34:57.565603484 +0000 UTC m=+1727.059466753" Dec 05 12:34:58 crc kubenswrapper[4807]: I1205 12:34:58.234914 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:34:58 crc kubenswrapper[4807]: E1205 12:34:58.235171 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:34:58 crc kubenswrapper[4807]: I1205 12:34:58.462379 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:34:58 crc kubenswrapper[4807]: I1205 12:34:58.690263 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:34:58 crc kubenswrapper[4807]: I1205 12:34:58.817495 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:34:59 crc kubenswrapper[4807]: I1205 12:34:59.717182 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:00 crc kubenswrapper[4807]: I1205 12:35:00.927345 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:02 crc kubenswrapper[4807]: I1205 12:35:02.187880 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:03 crc kubenswrapper[4807]: I1205 12:35:03.429307 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:03 crc kubenswrapper[4807]: I1205 12:35:03.690244 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:35:03 crc kubenswrapper[4807]: I1205 12:35:03.697201 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:35:03 crc kubenswrapper[4807]: I1205 12:35:03.817739 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:35:03 crc kubenswrapper[4807]: I1205 12:35:03.844182 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:35:03 crc kubenswrapper[4807]: I1205 12:35:03.893901 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:35:03 crc kubenswrapper[4807]: I1205 12:35:03.920893 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:35:04 crc kubenswrapper[4807]: I1205 12:35:04.226073 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:35:04 crc kubenswrapper[4807]: I1205 12:35:04.230951 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:35:04 crc kubenswrapper[4807]: I1205 12:35:04.254111 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:35:04 crc kubenswrapper[4807]: I1205 12:35:04.259272 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:35:04 crc kubenswrapper[4807]: I1205 12:35:04.641651 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:04 crc kubenswrapper[4807]: I1205 12:35:04.924500 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.294494 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-db-create-rmhlb"] Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.295596 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.303015 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-create-rmhlb"] Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.400820 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-828f-account-create-update-bwwwm"] Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.401864 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.404058 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-db-secret" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.409378 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4zg8\" (UniqueName: \"kubernetes.io/projected/efe7aa02-168c-490f-88d0-57de6b64f790-kube-api-access-k4zg8\") pod \"cinder-db-create-rmhlb\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.409465 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe7aa02-168c-490f-88d0-57de6b64f790-operator-scripts\") pod \"cinder-db-create-rmhlb\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.419734 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-828f-account-create-update-bwwwm"] Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.511032 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4zg8\" (UniqueName: \"kubernetes.io/projected/efe7aa02-168c-490f-88d0-57de6b64f790-kube-api-access-k4zg8\") pod \"cinder-db-create-rmhlb\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.511117 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe7aa02-168c-490f-88d0-57de6b64f790-operator-scripts\") pod \"cinder-db-create-rmhlb\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.511195 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-operator-scripts\") pod \"cinder-828f-account-create-update-bwwwm\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.511235 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g4lk\" (UniqueName: \"kubernetes.io/projected/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-kube-api-access-6g4lk\") pod \"cinder-828f-account-create-update-bwwwm\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.512069 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe7aa02-168c-490f-88d0-57de6b64f790-operator-scripts\") pod \"cinder-db-create-rmhlb\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.544707 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4zg8\" (UniqueName: \"kubernetes.io/projected/efe7aa02-168c-490f-88d0-57de6b64f790-kube-api-access-k4zg8\") pod \"cinder-db-create-rmhlb\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.612616 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-operator-scripts\") pod \"cinder-828f-account-create-update-bwwwm\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.612745 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g4lk\" (UniqueName: \"kubernetes.io/projected/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-kube-api-access-6g4lk\") pod \"cinder-828f-account-create-update-bwwwm\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.613493 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-operator-scripts\") pod \"cinder-828f-account-create-update-bwwwm\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.627447 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.632752 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g4lk\" (UniqueName: \"kubernetes.io/projected/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-kube-api-access-6g4lk\") pod \"cinder-828f-account-create-update-bwwwm\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:05 crc kubenswrapper[4807]: I1205 12:35:05.716713 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:06 crc kubenswrapper[4807]: W1205 12:35:06.246310 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefe7aa02_168c_490f_88d0_57de6b64f790.slice/crio-bdb5fa74624370672983385df73be219f2b251eeefcc8ca362b7aa152e075f21 WatchSource:0}: Error finding container bdb5fa74624370672983385df73be219f2b251eeefcc8ca362b7aa152e075f21: Status 404 returned error can't find the container with id bdb5fa74624370672983385df73be219f2b251eeefcc8ca362b7aa152e075f21 Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.246373 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-create-rmhlb"] Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.266068 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-create-rmhlb" event={"ID":"efe7aa02-168c-490f-88d0-57de6b64f790","Type":"ContainerStarted","Data":"bdb5fa74624370672983385df73be219f2b251eeefcc8ca362b7aa152e075f21"} Dec 05 12:35:06 crc kubenswrapper[4807]: W1205 12:35:06.434984 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2cd9d02_3e04_48f9_9c0d_21d36772ac95.slice/crio-0c0b598d01ff29935193f9c6e8df6eebdea167f9ab5b7ae069eeb09c7d148fca WatchSource:0}: Error finding container 0c0b598d01ff29935193f9c6e8df6eebdea167f9ab5b7ae069eeb09c7d148fca: Status 404 returned error can't find the container with id 0c0b598d01ff29935193f9c6e8df6eebdea167f9ab5b7ae069eeb09c7d148fca Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.435462 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-828f-account-create-update-bwwwm"] Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.602887 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.925265 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.925582 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-central-agent" containerID="cri-o://59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21" gracePeriod=30 Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.925627 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="proxy-httpd" containerID="cri-o://b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8" gracePeriod=30 Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.925719 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="sg-core" containerID="cri-o://c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4" gracePeriod=30 Dec 05 12:35:06 crc kubenswrapper[4807]: I1205 12:35:06.925768 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-notification-agent" containerID="cri-o://3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4" gracePeriod=30 Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.026129 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.184:3000/\": read tcp 10.217.0.2:48446->10.217.0.184:3000: read: connection reset by peer" Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.281187 4807 generic.go:334] "Generic (PLEG): container finished" podID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerID="b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8" exitCode=0 Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.281448 4807 generic.go:334] "Generic (PLEG): container finished" podID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerID="c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4" exitCode=2 Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.281460 4807 generic.go:334] "Generic (PLEG): container finished" podID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerID="59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21" exitCode=0 Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.281222 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerDied","Data":"b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8"} Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.281552 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerDied","Data":"c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4"} Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.281568 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerDied","Data":"59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21"} Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.282994 4807 generic.go:334] "Generic (PLEG): container finished" podID="d2cd9d02-3e04-48f9-9c0d-21d36772ac95" containerID="364c39668fdccbfb9483f968615480a26039292004744cd64cf4164299b11a29" exitCode=0 Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.283042 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" event={"ID":"d2cd9d02-3e04-48f9-9c0d-21d36772ac95","Type":"ContainerDied","Data":"364c39668fdccbfb9483f968615480a26039292004744cd64cf4164299b11a29"} Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.283057 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" event={"ID":"d2cd9d02-3e04-48f9-9c0d-21d36772ac95","Type":"ContainerStarted","Data":"0c0b598d01ff29935193f9c6e8df6eebdea167f9ab5b7ae069eeb09c7d148fca"} Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.284690 4807 generic.go:334] "Generic (PLEG): container finished" podID="efe7aa02-168c-490f-88d0-57de6b64f790" containerID="1576e052285f75fd620fba5a71d93132c29733de072e27420457f1ec7738ca95" exitCode=0 Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.284715 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-create-rmhlb" event={"ID":"efe7aa02-168c-490f-88d0-57de6b64f790","Type":"ContainerDied","Data":"1576e052285f75fd620fba5a71d93132c29733de072e27420457f1ec7738ca95"} Dec 05 12:35:07 crc kubenswrapper[4807]: I1205 12:35:07.818484 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.760699 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.772315 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.899234 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe7aa02-168c-490f-88d0-57de6b64f790-operator-scripts\") pod \"efe7aa02-168c-490f-88d0-57de6b64f790\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.899330 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g4lk\" (UniqueName: \"kubernetes.io/projected/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-kube-api-access-6g4lk\") pod \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.899396 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-operator-scripts\") pod \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\" (UID: \"d2cd9d02-3e04-48f9-9c0d-21d36772ac95\") " Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.899488 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4zg8\" (UniqueName: \"kubernetes.io/projected/efe7aa02-168c-490f-88d0-57de6b64f790-kube-api-access-k4zg8\") pod \"efe7aa02-168c-490f-88d0-57de6b64f790\" (UID: \"efe7aa02-168c-490f-88d0-57de6b64f790\") " Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.900151 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2cd9d02-3e04-48f9-9c0d-21d36772ac95" (UID: "d2cd9d02-3e04-48f9-9c0d-21d36772ac95"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.900548 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe7aa02-168c-490f-88d0-57de6b64f790-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "efe7aa02-168c-490f-88d0-57de6b64f790" (UID: "efe7aa02-168c-490f-88d0-57de6b64f790"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.905097 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-kube-api-access-6g4lk" (OuterVolumeSpecName: "kube-api-access-6g4lk") pod "d2cd9d02-3e04-48f9-9c0d-21d36772ac95" (UID: "d2cd9d02-3e04-48f9-9c0d-21d36772ac95"). InnerVolumeSpecName "kube-api-access-6g4lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:35:08 crc kubenswrapper[4807]: I1205 12:35:08.918802 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe7aa02-168c-490f-88d0-57de6b64f790-kube-api-access-k4zg8" (OuterVolumeSpecName: "kube-api-access-k4zg8") pod "efe7aa02-168c-490f-88d0-57de6b64f790" (UID: "efe7aa02-168c-490f-88d0-57de6b64f790"). InnerVolumeSpecName "kube-api-access-k4zg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.002035 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.002077 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4zg8\" (UniqueName: \"kubernetes.io/projected/efe7aa02-168c-490f-88d0-57de6b64f790-kube-api-access-k4zg8\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.002113 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe7aa02-168c-490f-88d0-57de6b64f790-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.002125 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g4lk\" (UniqueName: \"kubernetes.io/projected/d2cd9d02-3e04-48f9-9c0d-21d36772ac95-kube-api-access-6g4lk\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.052963 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.303177 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" event={"ID":"d2cd9d02-3e04-48f9-9c0d-21d36772ac95","Type":"ContainerDied","Data":"0c0b598d01ff29935193f9c6e8df6eebdea167f9ab5b7ae069eeb09c7d148fca"} Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.303479 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c0b598d01ff29935193f9c6e8df6eebdea167f9ab5b7ae069eeb09c7d148fca" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.303236 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-828f-account-create-update-bwwwm" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.305212 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-create-rmhlb" event={"ID":"efe7aa02-168c-490f-88d0-57de6b64f790","Type":"ContainerDied","Data":"bdb5fa74624370672983385df73be219f2b251eeefcc8ca362b7aa152e075f21"} Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.305254 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdb5fa74624370672983385df73be219f2b251eeefcc8ca362b7aa152e075f21" Dec 05 12:35:09 crc kubenswrapper[4807]: I1205 12:35:09.305302 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-rmhlb" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.302302 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.542062 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-db-sync-p5qc9"] Dec 05 12:35:10 crc kubenswrapper[4807]: E1205 12:35:10.542383 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe7aa02-168c-490f-88d0-57de6b64f790" containerName="mariadb-database-create" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.542400 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe7aa02-168c-490f-88d0-57de6b64f790" containerName="mariadb-database-create" Dec 05 12:35:10 crc kubenswrapper[4807]: E1205 12:35:10.542412 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cd9d02-3e04-48f9-9c0d-21d36772ac95" containerName="mariadb-account-create-update" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.542420 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cd9d02-3e04-48f9-9c0d-21d36772ac95" containerName="mariadb-account-create-update" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.542589 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe7aa02-168c-490f-88d0-57de6b64f790" containerName="mariadb-database-create" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.542604 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2cd9d02-3e04-48f9-9c0d-21d36772ac95" containerName="mariadb-account-create-update" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.543206 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.545710 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-config-data" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.547956 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-cinder-dockercfg-mjn4s" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.548263 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scripts" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.552268 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-p5qc9"] Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.627324 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-config-data\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.627378 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-scripts\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.627444 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-combined-ca-bundle\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.627475 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-db-sync-config-data\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.627500 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m84th\" (UniqueName: \"kubernetes.io/projected/03760c91-5719-41e3-a9d7-b23301e92be4-kube-api-access-m84th\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.627572 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03760c91-5719-41e3-a9d7-b23301e92be4-etc-machine-id\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.729648 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03760c91-5719-41e3-a9d7-b23301e92be4-etc-machine-id\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.729772 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-config-data\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.729796 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-scripts\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.729858 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-combined-ca-bundle\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.729887 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-db-sync-config-data\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.729913 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m84th\" (UniqueName: \"kubernetes.io/projected/03760c91-5719-41e3-a9d7-b23301e92be4-kube-api-access-m84th\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.730291 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03760c91-5719-41e3-a9d7-b23301e92be4-etc-machine-id\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.737477 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-combined-ca-bundle\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.739215 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-scripts\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.739455 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-db-sync-config-data\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.739764 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-config-data\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.751693 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m84th\" (UniqueName: \"kubernetes.io/projected/03760c91-5719-41e3-a9d7-b23301e92be4-kube-api-access-m84th\") pod \"cinder-db-sync-p5qc9\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:10 crc kubenswrapper[4807]: I1205 12:35:10.859864 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:11 crc kubenswrapper[4807]: I1205 12:35:11.373569 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-p5qc9"] Dec 05 12:35:11 crc kubenswrapper[4807]: I1205 12:35:11.500837 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.332921 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.334894 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" event={"ID":"03760c91-5719-41e3-a9d7-b23301e92be4","Type":"ContainerStarted","Data":"9d7d8e8212727b61e9d45c80e2c12a8600a1e5b45bea8d9e3cfe4a66691c1eea"} Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.339185 4807 generic.go:334] "Generic (PLEG): container finished" podID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerID="3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4" exitCode=0 Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.339239 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.339284 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerDied","Data":"3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4"} Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.339559 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5cd50e15-557b-4d86-af33-56ce3ace1698","Type":"ContainerDied","Data":"a7a19177af34416b46342ea9dd7c388c576e32ae643988561841577df9ddb562"} Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.339634 4807 scope.go:117] "RemoveContainer" containerID="b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.392635 4807 scope.go:117] "RemoveContainer" containerID="c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.411786 4807 scope.go:117] "RemoveContainer" containerID="3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.438221 4807 scope.go:117] "RemoveContainer" containerID="59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.464758 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-combined-ca-bundle\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.465633 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-sg-core-conf-yaml\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.466035 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-log-httpd\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.466117 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-run-httpd\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.466179 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-scripts\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.466312 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j8ck\" (UniqueName: \"kubernetes.io/projected/5cd50e15-557b-4d86-af33-56ce3ace1698-kube-api-access-9j8ck\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.466396 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-ceilometer-tls-certs\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.466503 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-config-data\") pod \"5cd50e15-557b-4d86-af33-56ce3ace1698\" (UID: \"5cd50e15-557b-4d86-af33-56ce3ace1698\") " Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.467159 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.467773 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.470955 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-scripts" (OuterVolumeSpecName: "scripts") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.471582 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd50e15-557b-4d86-af33-56ce3ace1698-kube-api-access-9j8ck" (OuterVolumeSpecName: "kube-api-access-9j8ck") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "kube-api-access-9j8ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.481573 4807 scope.go:117] "RemoveContainer" containerID="b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8" Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.482439 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8\": container with ID starting with b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8 not found: ID does not exist" containerID="b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.482504 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8"} err="failed to get container status \"b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8\": rpc error: code = NotFound desc = could not find container \"b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8\": container with ID starting with b10481507f9b6948ea54e89428ef90c5c28f0d8c7a80a6ad4fbd48f08e868ee8 not found: ID does not exist" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.482553 4807 scope.go:117] "RemoveContainer" containerID="c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4" Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.483128 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4\": container with ID starting with c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4 not found: ID does not exist" containerID="c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.483177 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4"} err="failed to get container status \"c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4\": rpc error: code = NotFound desc = could not find container \"c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4\": container with ID starting with c324b31a556c1c71cea9b2338f44b09dd31e90e6b2a8062468c5c81bca5566f4 not found: ID does not exist" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.483207 4807 scope.go:117] "RemoveContainer" containerID="3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4" Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.484148 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4\": container with ID starting with 3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4 not found: ID does not exist" containerID="3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.484189 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4"} err="failed to get container status \"3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4\": rpc error: code = NotFound desc = could not find container \"3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4\": container with ID starting with 3aae97c9084312086e69029931c4a132a19cf5c7e2156071f28ce328a0fed6c4 not found: ID does not exist" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.484215 4807 scope.go:117] "RemoveContainer" containerID="59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21" Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.484602 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21\": container with ID starting with 59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21 not found: ID does not exist" containerID="59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.484633 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21"} err="failed to get container status \"59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21\": rpc error: code = NotFound desc = could not find container \"59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21\": container with ID starting with 59a18426a73eb38ceace92b5b0e373a70d873a5c9dd8588d41038d8a39ebac21 not found: ID does not exist" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.500350 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.521203 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.541609 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.569180 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.569216 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.569227 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cd50e15-557b-4d86-af33-56ce3ace1698-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.569241 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.569251 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j8ck\" (UniqueName: \"kubernetes.io/projected/5cd50e15-557b-4d86-af33-56ce3ace1698-kube-api-access-9j8ck\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.569266 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.569276 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.574764 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-config-data" (OuterVolumeSpecName: "config-data") pod "5cd50e15-557b-4d86-af33-56ce3ace1698" (UID: "5cd50e15-557b-4d86-af33-56ce3ace1698"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.670247 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.670781 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd50e15-557b-4d86-af33-56ce3ace1698-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.678916 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.706269 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.706677 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-notification-agent" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.706700 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-notification-agent" Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.706726 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-central-agent" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.706735 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-central-agent" Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.706750 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="proxy-httpd" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.706757 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="proxy-httpd" Dec 05 12:35:12 crc kubenswrapper[4807]: E1205 12:35:12.706772 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="sg-core" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.706780 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="sg-core" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.706984 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="sg-core" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.707010 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="proxy-httpd" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.707027 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-central-agent" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.707041 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" containerName="ceilometer-notification-agent" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.709835 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.712170 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.714366 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.714738 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.718927 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.771372 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874114 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krgkq\" (UniqueName: \"kubernetes.io/projected/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-kube-api-access-krgkq\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874188 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874330 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874391 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-log-httpd\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874439 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874474 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-scripts\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874513 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-run-httpd\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.874560 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-config-data\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.975774 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-config-data\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.975854 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krgkq\" (UniqueName: \"kubernetes.io/projected/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-kube-api-access-krgkq\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.975884 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.975963 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.976011 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-log-httpd\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.976042 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.976071 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-scripts\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.976101 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-run-httpd\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.977061 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-log-httpd\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.977134 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-run-httpd\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.981659 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-config-data\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.981821 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-scripts\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.981916 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.983294 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.995718 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:12 crc kubenswrapper[4807]: I1205 12:35:12.996201 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krgkq\" (UniqueName: \"kubernetes.io/projected/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-kube-api-access-krgkq\") pod \"ceilometer-0\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:13 crc kubenswrapper[4807]: I1205 12:35:13.041019 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:13 crc kubenswrapper[4807]: I1205 12:35:13.236204 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:35:13 crc kubenswrapper[4807]: E1205 12:35:13.236409 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:35:13 crc kubenswrapper[4807]: I1205 12:35:13.256731 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd50e15-557b-4d86-af33-56ce3ace1698" path="/var/lib/kubelet/pods/5cd50e15-557b-4d86-af33-56ce3ace1698/volumes" Dec 05 12:35:13 crc kubenswrapper[4807]: I1205 12:35:13.531108 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:35:13 crc kubenswrapper[4807]: W1205 12:35:13.543360 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57a61ee2_5709_4a11_b12f_e1b6a9c641f6.slice/crio-eb657ddc78fe52bdcaed024debac3dff4153a3c1f713eac39992c6b743ac7c2c WatchSource:0}: Error finding container eb657ddc78fe52bdcaed024debac3dff4153a3c1f713eac39992c6b743ac7c2c: Status 404 returned error can't find the container with id eb657ddc78fe52bdcaed024debac3dff4153a3c1f713eac39992c6b743ac7c2c Dec 05 12:35:14 crc kubenswrapper[4807]: I1205 12:35:14.053155 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:14 crc kubenswrapper[4807]: I1205 12:35:14.361027 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerStarted","Data":"eb657ddc78fe52bdcaed024debac3dff4153a3c1f713eac39992c6b743ac7c2c"} Dec 05 12:35:15 crc kubenswrapper[4807]: I1205 12:35:15.292541 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:15 crc kubenswrapper[4807]: I1205 12:35:15.374183 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerStarted","Data":"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d"} Dec 05 12:35:15 crc kubenswrapper[4807]: I1205 12:35:15.374236 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerStarted","Data":"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4"} Dec 05 12:35:16 crc kubenswrapper[4807]: I1205 12:35:16.385966 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerStarted","Data":"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f"} Dec 05 12:35:16 crc kubenswrapper[4807]: I1205 12:35:16.577416 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:17 crc kubenswrapper[4807]: I1205 12:35:17.892339 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:18 crc kubenswrapper[4807]: I1205 12:35:18.407454 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerStarted","Data":"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12"} Dec 05 12:35:18 crc kubenswrapper[4807]: I1205 12:35:18.407707 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:18 crc kubenswrapper[4807]: I1205 12:35:18.445496 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.502948771 podStartE2EDuration="6.445466096s" podCreationTimestamp="2025-12-05 12:35:12 +0000 UTC" firstStartedPulling="2025-12-05 12:35:13.554449618 +0000 UTC m=+1743.048312887" lastFinishedPulling="2025-12-05 12:35:17.496966943 +0000 UTC m=+1746.990830212" observedRunningTime="2025-12-05 12:35:18.44480923 +0000 UTC m=+1747.938672509" watchObservedRunningTime="2025-12-05 12:35:18.445466096 +0000 UTC m=+1747.939329365" Dec 05 12:35:19 crc kubenswrapper[4807]: I1205 12:35:19.162099 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:20 crc kubenswrapper[4807]: I1205 12:35:20.461250 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:21 crc kubenswrapper[4807]: I1205 12:35:21.704948 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:23 crc kubenswrapper[4807]: I1205 12:35:23.008441 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:24 crc kubenswrapper[4807]: I1205 12:35:24.255608 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:25 crc kubenswrapper[4807]: I1205 12:35:25.236575 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:35:25 crc kubenswrapper[4807]: E1205 12:35:25.236810 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:35:25 crc kubenswrapper[4807]: I1205 12:35:25.540566 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:26 crc kubenswrapper[4807]: I1205 12:35:26.768646 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:28 crc kubenswrapper[4807]: I1205 12:35:28.003781 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:28 crc kubenswrapper[4807]: E1205 12:35:28.097411 4807 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 05 12:35:28 crc kubenswrapper[4807]: E1205 12:35:28.097632 4807 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m84th,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-p5qc9_watcher-kuttl-default(03760c91-5719-41e3-a9d7-b23301e92be4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 12:35:28 crc kubenswrapper[4807]: E1205 12:35:28.098749 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" podUID="03760c91-5719-41e3-a9d7-b23301e92be4" Dec 05 12:35:28 crc kubenswrapper[4807]: E1205 12:35:28.516898 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" podUID="03760c91-5719-41e3-a9d7-b23301e92be4" Dec 05 12:35:29 crc kubenswrapper[4807]: I1205 12:35:29.328674 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:30 crc kubenswrapper[4807]: I1205 12:35:30.590742 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:31 crc kubenswrapper[4807]: I1205 12:35:31.841334 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:33 crc kubenswrapper[4807]: I1205 12:35:33.094054 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:34 crc kubenswrapper[4807]: I1205 12:35:34.358494 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:35 crc kubenswrapper[4807]: I1205 12:35:35.571173 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:36 crc kubenswrapper[4807]: I1205 12:35:36.795045 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:38 crc kubenswrapper[4807]: I1205 12:35:38.056369 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:38 crc kubenswrapper[4807]: I1205 12:35:38.235729 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:35:38 crc kubenswrapper[4807]: E1205 12:35:38.236070 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:35:39 crc kubenswrapper[4807]: I1205 12:35:39.308627 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:40 crc kubenswrapper[4807]: I1205 12:35:40.537581 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:41 crc kubenswrapper[4807]: I1205 12:35:41.630546 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" event={"ID":"03760c91-5719-41e3-a9d7-b23301e92be4","Type":"ContainerStarted","Data":"2440865f212d0a177000e6a29ebc554fb0a75168bca253fb04e424489c5a4e24"} Dec 05 12:35:41 crc kubenswrapper[4807]: I1205 12:35:41.656569 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" podStartSLOduration=2.733613118 podStartE2EDuration="31.656544462s" podCreationTimestamp="2025-12-05 12:35:10 +0000 UTC" firstStartedPulling="2025-12-05 12:35:11.383144269 +0000 UTC m=+1740.877007538" lastFinishedPulling="2025-12-05 12:35:40.306075623 +0000 UTC m=+1769.799938882" observedRunningTime="2025-12-05 12:35:41.656232544 +0000 UTC m=+1771.150095803" watchObservedRunningTime="2025-12-05 12:35:41.656544462 +0000 UTC m=+1771.150407731" Dec 05 12:35:41 crc kubenswrapper[4807]: I1205 12:35:41.799771 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:43 crc kubenswrapper[4807]: I1205 12:35:43.047924 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:35:43 crc kubenswrapper[4807]: I1205 12:35:43.052285 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:44 crc kubenswrapper[4807]: I1205 12:35:44.314660 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:45 crc kubenswrapper[4807]: I1205 12:35:45.559167 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:45 crc kubenswrapper[4807]: I1205 12:35:45.661940 4807 generic.go:334] "Generic (PLEG): container finished" podID="03760c91-5719-41e3-a9d7-b23301e92be4" containerID="2440865f212d0a177000e6a29ebc554fb0a75168bca253fb04e424489c5a4e24" exitCode=0 Dec 05 12:35:45 crc kubenswrapper[4807]: I1205 12:35:45.661998 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" event={"ID":"03760c91-5719-41e3-a9d7-b23301e92be4","Type":"ContainerDied","Data":"2440865f212d0a177000e6a29ebc554fb0a75168bca253fb04e424489c5a4e24"} Dec 05 12:35:46 crc kubenswrapper[4807]: I1205 12:35:46.767020 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.012584 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.171195 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m84th\" (UniqueName: \"kubernetes.io/projected/03760c91-5719-41e3-a9d7-b23301e92be4-kube-api-access-m84th\") pod \"03760c91-5719-41e3-a9d7-b23301e92be4\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.171711 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-scripts\") pod \"03760c91-5719-41e3-a9d7-b23301e92be4\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.171759 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03760c91-5719-41e3-a9d7-b23301e92be4-etc-machine-id\") pod \"03760c91-5719-41e3-a9d7-b23301e92be4\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.171798 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-combined-ca-bundle\") pod \"03760c91-5719-41e3-a9d7-b23301e92be4\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.171851 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-db-sync-config-data\") pod \"03760c91-5719-41e3-a9d7-b23301e92be4\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.171892 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-config-data\") pod \"03760c91-5719-41e3-a9d7-b23301e92be4\" (UID: \"03760c91-5719-41e3-a9d7-b23301e92be4\") " Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.171891 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03760c91-5719-41e3-a9d7-b23301e92be4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "03760c91-5719-41e3-a9d7-b23301e92be4" (UID: "03760c91-5719-41e3-a9d7-b23301e92be4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.172336 4807 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03760c91-5719-41e3-a9d7-b23301e92be4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.176679 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "03760c91-5719-41e3-a9d7-b23301e92be4" (UID: "03760c91-5719-41e3-a9d7-b23301e92be4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.176975 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03760c91-5719-41e3-a9d7-b23301e92be4-kube-api-access-m84th" (OuterVolumeSpecName: "kube-api-access-m84th") pod "03760c91-5719-41e3-a9d7-b23301e92be4" (UID: "03760c91-5719-41e3-a9d7-b23301e92be4"). InnerVolumeSpecName "kube-api-access-m84th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.176989 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-scripts" (OuterVolumeSpecName: "scripts") pod "03760c91-5719-41e3-a9d7-b23301e92be4" (UID: "03760c91-5719-41e3-a9d7-b23301e92be4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.200902 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03760c91-5719-41e3-a9d7-b23301e92be4" (UID: "03760c91-5719-41e3-a9d7-b23301e92be4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.219441 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-config-data" (OuterVolumeSpecName: "config-data") pod "03760c91-5719-41e3-a9d7-b23301e92be4" (UID: "03760c91-5719-41e3-a9d7-b23301e92be4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.273495 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m84th\" (UniqueName: \"kubernetes.io/projected/03760c91-5719-41e3-a9d7-b23301e92be4-kube-api-access-m84th\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.273566 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.273579 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.273588 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.273596 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03760c91-5719-41e3-a9d7-b23301e92be4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.679130 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.679044 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-p5qc9" event={"ID":"03760c91-5719-41e3-a9d7-b23301e92be4","Type":"ContainerDied","Data":"9d7d8e8212727b61e9d45c80e2c12a8600a1e5b45bea8d9e3cfe4a66691c1eea"} Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.679279 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d7d8e8212727b61e9d45c80e2c12a8600a1e5b45bea8d9e3cfe4a66691c1eea" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.997494 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:35:47 crc kubenswrapper[4807]: E1205 12:35:47.997883 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03760c91-5719-41e3-a9d7-b23301e92be4" containerName="cinder-db-sync" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.997896 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="03760c91-5719-41e3-a9d7-b23301e92be4" containerName="cinder-db-sync" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.998055 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="03760c91-5719-41e3-a9d7-b23301e92be4" containerName="cinder-db-sync" Dec 05 12:35:47 crc kubenswrapper[4807]: I1205 12:35:47.998956 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.005621 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scheduler-config-data" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.005727 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-config-data" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.005823 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-cinder-dockercfg-mjn4s" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.006230 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scripts" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.007503 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.030987 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.062426 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.063948 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.067143 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-backup-config-data" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.069827 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.085712 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.085758 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c28b308-84a6-423f-8b24-9814aa6e275b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.085817 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.085860 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.085931 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-scripts\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.085991 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqprb\" (UniqueName: \"kubernetes.io/projected/8c28b308-84a6-423f-8b24-9814aa6e275b-kube-api-access-jqprb\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.086019 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.187674 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-dev\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.187712 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.187739 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqprb\" (UniqueName: \"kubernetes.io/projected/8c28b308-84a6-423f-8b24-9814aa6e275b-kube-api-access-jqprb\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.187762 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-run\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.187926 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.187975 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188026 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfjb2\" (UniqueName: \"kubernetes.io/projected/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-kube-api-access-nfjb2\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188072 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188148 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188209 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188254 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-scripts\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188277 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-lib-modules\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188300 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188337 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c28b308-84a6-423f-8b24-9814aa6e275b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188382 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-sys\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188434 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c28b308-84a6-423f-8b24-9814aa6e275b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188452 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188502 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188555 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data-custom\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188610 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188664 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-nvme\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188681 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-scripts\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188734 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.188779 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.191297 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.191727 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-scripts\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.192352 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.194920 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.198005 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.220006 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqprb\" (UniqueName: \"kubernetes.io/projected/8c28b308-84a6-423f-8b24-9814aa6e275b-kube-api-access-jqprb\") pod \"cinder-scheduler-0\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290146 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290430 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290458 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-dev\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290475 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290490 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-run\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290511 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290548 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfjb2\" (UniqueName: \"kubernetes.io/projected/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-kube-api-access-nfjb2\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290569 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290591 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290617 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290635 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-scripts\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290650 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-lib-modules\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290677 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-sys\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290705 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290724 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data-custom\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290761 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-nvme\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290899 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-nvme\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.290949 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.291136 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.291164 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-dev\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.291183 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.291204 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-run\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.291753 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.291819 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-sys\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.291834 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-lib-modules\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.292178 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.293570 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.294970 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.296360 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data-custom\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.298166 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.302905 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-scripts\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.303288 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-api-config-data" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.303787 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.311035 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.315236 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfjb2\" (UniqueName: \"kubernetes.io/projected/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-kube-api-access-nfjb2\") pod \"cinder-backup-0\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.323379 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.324166 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.391809 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx6ck\" (UniqueName: \"kubernetes.io/projected/d081f1da-3e50-4ace-b199-f72235d17102-kube-api-access-qx6ck\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.392024 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-scripts\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.392089 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.392130 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data-custom\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.392148 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.392171 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.392219 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d081f1da-3e50-4ace-b199-f72235d17102-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.392244 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d081f1da-3e50-4ace-b199-f72235d17102-logs\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.403696 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494668 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx6ck\" (UniqueName: \"kubernetes.io/projected/d081f1da-3e50-4ace-b199-f72235d17102-kube-api-access-qx6ck\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494752 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-scripts\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494820 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494863 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data-custom\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494882 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494904 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494948 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d081f1da-3e50-4ace-b199-f72235d17102-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.494976 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d081f1da-3e50-4ace-b199-f72235d17102-logs\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.495562 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d081f1da-3e50-4ace-b199-f72235d17102-logs\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.498372 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d081f1da-3e50-4ace-b199-f72235d17102-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.507495 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.509250 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-scripts\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.509948 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.510077 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.514201 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data-custom\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.560267 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx6ck\" (UniqueName: \"kubernetes.io/projected/d081f1da-3e50-4ace-b199-f72235d17102-kube-api-access-qx6ck\") pod \"cinder-api-0\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.730884 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:48 crc kubenswrapper[4807]: I1205 12:35:48.896450 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:35:48 crc kubenswrapper[4807]: W1205 12:35:48.908268 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c28b308_84a6_423f_8b24_9814aa6e275b.slice/crio-1edf90b737b16307418426d712e140c068986b673c8e88f04cf2d8c3c1791b60 WatchSource:0}: Error finding container 1edf90b737b16307418426d712e140c068986b673c8e88f04cf2d8c3c1791b60: Status 404 returned error can't find the container with id 1edf90b737b16307418426d712e140c068986b673c8e88f04cf2d8c3c1791b60 Dec 05 12:35:49 crc kubenswrapper[4807]: I1205 12:35:49.150935 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:35:49 crc kubenswrapper[4807]: W1205 12:35:49.155845 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91cfacd9_7fa0_4e43_8a2e_560c74e1a865.slice/crio-cbd8b358f84bc5f2b29b4f11a85977956616165149eb7ae0954374cc94976364 WatchSource:0}: Error finding container cbd8b358f84bc5f2b29b4f11a85977956616165149eb7ae0954374cc94976364: Status 404 returned error can't find the container with id cbd8b358f84bc5f2b29b4f11a85977956616165149eb7ae0954374cc94976364 Dec 05 12:35:49 crc kubenswrapper[4807]: I1205 12:35:49.235300 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:35:49 crc kubenswrapper[4807]: E1205 12:35:49.235732 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:35:49 crc kubenswrapper[4807]: I1205 12:35:49.278925 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:49 crc kubenswrapper[4807]: I1205 12:35:49.302550 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:49 crc kubenswrapper[4807]: I1205 12:35:49.722404 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"91cfacd9-7fa0-4e43-8a2e-560c74e1a865","Type":"ContainerStarted","Data":"cbd8b358f84bc5f2b29b4f11a85977956616165149eb7ae0954374cc94976364"} Dec 05 12:35:49 crc kubenswrapper[4807]: I1205 12:35:49.728403 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"8c28b308-84a6-423f-8b24-9814aa6e275b","Type":"ContainerStarted","Data":"1edf90b737b16307418426d712e140c068986b673c8e88f04cf2d8c3c1791b60"} Dec 05 12:35:49 crc kubenswrapper[4807]: I1205 12:35:49.730548 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d081f1da-3e50-4ace-b199-f72235d17102","Type":"ContainerStarted","Data":"707dddfd65bbebe4a21e5115810cfdacd049e5d0ebf7a54b4c7f6b09db07e8c2"} Dec 05 12:35:50 crc kubenswrapper[4807]: I1205 12:35:50.346002 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:50 crc kubenswrapper[4807]: I1205 12:35:50.496077 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:50 crc kubenswrapper[4807]: I1205 12:35:50.759989 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d081f1da-3e50-4ace-b199-f72235d17102","Type":"ContainerStarted","Data":"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65"} Dec 05 12:35:51 crc kubenswrapper[4807]: I1205 12:35:51.771742 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:51 crc kubenswrapper[4807]: I1205 12:35:51.775195 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"91cfacd9-7fa0-4e43-8a2e-560c74e1a865","Type":"ContainerStarted","Data":"3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b"} Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.785281 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"8c28b308-84a6-423f-8b24-9814aa6e275b","Type":"ContainerStarted","Data":"57a9bf66514db280695b8253af4214896cd847bc9afc6b711db29158d5e4c62d"} Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.786725 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"8c28b308-84a6-423f-8b24-9814aa6e275b","Type":"ContainerStarted","Data":"b440b50270ec6b4666fdf3288b1e1d3171804a477605a28866a45e40416679f3"} Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.787990 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d081f1da-3e50-4ace-b199-f72235d17102","Type":"ContainerStarted","Data":"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c"} Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.788137 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api-log" containerID="cri-o://769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65" gracePeriod=30 Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.788382 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.788415 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api" containerID="cri-o://990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c" gracePeriod=30 Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.794369 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"91cfacd9-7fa0-4e43-8a2e-560c74e1a865","Type":"ContainerStarted","Data":"b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2"} Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.819959 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-scheduler-0" podStartSLOduration=4.000624659 podStartE2EDuration="5.819935997s" podCreationTimestamp="2025-12-05 12:35:47 +0000 UTC" firstStartedPulling="2025-12-05 12:35:48.916126839 +0000 UTC m=+1778.409990108" lastFinishedPulling="2025-12-05 12:35:50.735438177 +0000 UTC m=+1780.229301446" observedRunningTime="2025-12-05 12:35:52.81346079 +0000 UTC m=+1782.307324059" watchObservedRunningTime="2025-12-05 12:35:52.819935997 +0000 UTC m=+1782.313799296" Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.856145 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-api-0" podStartSLOduration=4.856126859 podStartE2EDuration="4.856126859s" podCreationTimestamp="2025-12-05 12:35:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:35:52.840090309 +0000 UTC m=+1782.333953578" watchObservedRunningTime="2025-12-05 12:35:52.856126859 +0000 UTC m=+1782.349990128" Dec 05 12:35:52 crc kubenswrapper[4807]: I1205 12:35:52.875047 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-backup-0" podStartSLOduration=3.296323027 podStartE2EDuration="4.87502842s" podCreationTimestamp="2025-12-05 12:35:48 +0000 UTC" firstStartedPulling="2025-12-05 12:35:49.158123378 +0000 UTC m=+1778.651986647" lastFinishedPulling="2025-12-05 12:35:50.736828771 +0000 UTC m=+1780.230692040" observedRunningTime="2025-12-05 12:35:52.866721188 +0000 UTC m=+1782.360584477" watchObservedRunningTime="2025-12-05 12:35:52.87502842 +0000 UTC m=+1782.368891679" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.040109 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.325932 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.404655 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.513960 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.611580 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data-custom\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.611711 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d081f1da-3e50-4ace-b199-f72235d17102-etc-machine-id\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.611773 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-combined-ca-bundle\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.611830 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d081f1da-3e50-4ace-b199-f72235d17102-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.611850 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.611929 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d081f1da-3e50-4ace-b199-f72235d17102-logs\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.611974 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-scripts\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.612049 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx6ck\" (UniqueName: \"kubernetes.io/projected/d081f1da-3e50-4ace-b199-f72235d17102-kube-api-access-qx6ck\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.612093 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-cert-memcached-mtls\") pod \"d081f1da-3e50-4ace-b199-f72235d17102\" (UID: \"d081f1da-3e50-4ace-b199-f72235d17102\") " Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.612548 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d081f1da-3e50-4ace-b199-f72235d17102-logs" (OuterVolumeSpecName: "logs") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.613711 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d081f1da-3e50-4ace-b199-f72235d17102-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.613736 4807 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d081f1da-3e50-4ace-b199-f72235d17102-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.617265 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.634260 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-scripts" (OuterVolumeSpecName: "scripts") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.666788 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d081f1da-3e50-4ace-b199-f72235d17102-kube-api-access-qx6ck" (OuterVolumeSpecName: "kube-api-access-qx6ck") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "kube-api-access-qx6ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.690683 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data" (OuterVolumeSpecName: "config-data") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.708052 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.715186 4807 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.715222 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.715237 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.715251 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx6ck\" (UniqueName: \"kubernetes.io/projected/d081f1da-3e50-4ace-b199-f72235d17102-kube-api-access-qx6ck\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.715264 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.716637 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d081f1da-3e50-4ace-b199-f72235d17102" (UID: "d081f1da-3e50-4ace-b199-f72235d17102"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.805332 4807 generic.go:334] "Generic (PLEG): container finished" podID="d081f1da-3e50-4ace-b199-f72235d17102" containerID="990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c" exitCode=0 Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.805575 4807 generic.go:334] "Generic (PLEG): container finished" podID="d081f1da-3e50-4ace-b199-f72235d17102" containerID="769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65" exitCode=143 Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.805448 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.805473 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d081f1da-3e50-4ace-b199-f72235d17102","Type":"ContainerDied","Data":"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c"} Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.805815 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d081f1da-3e50-4ace-b199-f72235d17102","Type":"ContainerDied","Data":"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65"} Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.805925 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d081f1da-3e50-4ace-b199-f72235d17102","Type":"ContainerDied","Data":"707dddfd65bbebe4a21e5115810cfdacd049e5d0ebf7a54b4c7f6b09db07e8c2"} Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.805988 4807 scope.go:117] "RemoveContainer" containerID="990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.817660 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d081f1da-3e50-4ace-b199-f72235d17102-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.852597 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.854898 4807 scope.go:117] "RemoveContainer" containerID="769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.869435 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.875347 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:53 crc kubenswrapper[4807]: E1205 12:35:53.875773 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.875796 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api" Dec 05 12:35:53 crc kubenswrapper[4807]: E1205 12:35:53.875820 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api-log" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.875828 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api-log" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.876023 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.876039 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d081f1da-3e50-4ace-b199-f72235d17102" containerName="cinder-api-log" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.877140 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.882731 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-api-config-data" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.882985 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-cinder-internal-svc" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.883330 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-cinder-public-svc" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.891197 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.923308 4807 scope.go:117] "RemoveContainer" containerID="990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c" Dec 05 12:35:53 crc kubenswrapper[4807]: E1205 12:35:53.940032 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c\": container with ID starting with 990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c not found: ID does not exist" containerID="990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.940095 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c"} err="failed to get container status \"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c\": rpc error: code = NotFound desc = could not find container \"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c\": container with ID starting with 990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c not found: ID does not exist" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.940132 4807 scope.go:117] "RemoveContainer" containerID="769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65" Dec 05 12:35:53 crc kubenswrapper[4807]: E1205 12:35:53.946711 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65\": container with ID starting with 769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65 not found: ID does not exist" containerID="769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.946766 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65"} err="failed to get container status \"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65\": rpc error: code = NotFound desc = could not find container \"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65\": container with ID starting with 769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65 not found: ID does not exist" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.946796 4807 scope.go:117] "RemoveContainer" containerID="990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.950655 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c"} err="failed to get container status \"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c\": rpc error: code = NotFound desc = could not find container \"990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c\": container with ID starting with 990cc1e7d9437e2ceee97eb1505eb3cfc1df77fa66c311c95b15b66ca0a6b56c not found: ID does not exist" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.950699 4807 scope.go:117] "RemoveContainer" containerID="769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65" Dec 05 12:35:53 crc kubenswrapper[4807]: I1205 12:35:53.953766 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65"} err="failed to get container status \"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65\": rpc error: code = NotFound desc = could not find container \"769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65\": container with ID starting with 769c5618b79f2f6c24b1f3e8956b060cdd5200b797ca1405fcb1fcfdf59eac65 not found: ID does not exist" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021150 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021197 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021326 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-scripts\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021400 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data-custom\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021462 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63b5b13b-b94f-4208-acda-836de99625f0-logs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021647 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021679 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021705 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63b5b13b-b94f-4208-acda-836de99625f0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021731 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.021757 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/63b5b13b-b94f-4208-acda-836de99625f0-kube-api-access-kv8z9\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.122856 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.122905 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.122927 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-scripts\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.122941 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data-custom\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.122969 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63b5b13b-b94f-4208-acda-836de99625f0-logs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.123012 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.123027 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.123047 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63b5b13b-b94f-4208-acda-836de99625f0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.123070 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.123150 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63b5b13b-b94f-4208-acda-836de99625f0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.123251 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/63b5b13b-b94f-4208-acda-836de99625f0-kube-api-access-kv8z9\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.124002 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63b5b13b-b94f-4208-acda-836de99625f0-logs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.128110 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.129518 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.129972 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-scripts\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.130051 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.130161 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.134173 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.134243 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data-custom\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.143976 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/63b5b13b-b94f-4208-acda-836de99625f0-kube-api-access-kv8z9\") pod \"cinder-api-0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.206806 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.308227 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.760017 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:35:54 crc kubenswrapper[4807]: I1205 12:35:54.816378 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"63b5b13b-b94f-4208-acda-836de99625f0","Type":"ContainerStarted","Data":"a5f7a7c52872218077fffdf36f172f61574142836ff61ac203552c47889d28eb"} Dec 05 12:35:55 crc kubenswrapper[4807]: I1205 12:35:55.246665 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d081f1da-3e50-4ace-b199-f72235d17102" path="/var/lib/kubelet/pods/d081f1da-3e50-4ace-b199-f72235d17102/volumes" Dec 05 12:35:55 crc kubenswrapper[4807]: I1205 12:35:55.546563 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:55 crc kubenswrapper[4807]: I1205 12:35:55.830133 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"63b5b13b-b94f-4208-acda-836de99625f0","Type":"ContainerStarted","Data":"8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131"} Dec 05 12:35:56 crc kubenswrapper[4807]: I1205 12:35:56.763557 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:56 crc kubenswrapper[4807]: I1205 12:35:56.841551 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"63b5b13b-b94f-4208-acda-836de99625f0","Type":"ContainerStarted","Data":"a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0"} Dec 05 12:35:56 crc kubenswrapper[4807]: I1205 12:35:56.841785 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:35:56 crc kubenswrapper[4807]: I1205 12:35:56.861489 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-api-0" podStartSLOduration=3.861469392 podStartE2EDuration="3.861469392s" podCreationTimestamp="2025-12-05 12:35:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:35:56.86013797 +0000 UTC m=+1786.354001259" watchObservedRunningTime="2025-12-05 12:35:56.861469392 +0000 UTC m=+1786.355332661" Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.073303 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.661746 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.685425 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.748454 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.793263 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.856771 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="cinder-scheduler" containerID="cri-o://b440b50270ec6b4666fdf3288b1e1d3171804a477605a28866a45e40416679f3" gracePeriod=30 Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.856818 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="probe" containerID="cri-o://57a9bf66514db280695b8253af4214896cd847bc9afc6b711db29158d5e4c62d" gracePeriod=30 Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.856926 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="cinder-backup" containerID="cri-o://3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b" gracePeriod=30 Dec 05 12:35:58 crc kubenswrapper[4807]: I1205 12:35:58.856927 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="probe" containerID="cri-o://b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2" gracePeriod=30 Dec 05 12:35:59 crc kubenswrapper[4807]: I1205 12:35:59.311328 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:35:59 crc kubenswrapper[4807]: I1205 12:35:59.866448 4807 generic.go:334] "Generic (PLEG): container finished" podID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerID="b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2" exitCode=0 Dec 05 12:35:59 crc kubenswrapper[4807]: I1205 12:35:59.866554 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"91cfacd9-7fa0-4e43-8a2e-560c74e1a865","Type":"ContainerDied","Data":"b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2"} Dec 05 12:35:59 crc kubenswrapper[4807]: I1205 12:35:59.868502 4807 generic.go:334] "Generic (PLEG): container finished" podID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerID="57a9bf66514db280695b8253af4214896cd847bc9afc6b711db29158d5e4c62d" exitCode=0 Dec 05 12:35:59 crc kubenswrapper[4807]: I1205 12:35:59.868550 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"8c28b308-84a6-423f-8b24-9814aa6e275b","Type":"ContainerDied","Data":"57a9bf66514db280695b8253af4214896cd847bc9afc6b711db29158d5e4c62d"} Dec 05 12:36:00 crc kubenswrapper[4807]: I1205 12:36:00.542359 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:36:00 crc kubenswrapper[4807]: I1205 12:36:00.879881 4807 generic.go:334] "Generic (PLEG): container finished" podID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerID="b440b50270ec6b4666fdf3288b1e1d3171804a477605a28866a45e40416679f3" exitCode=0 Dec 05 12:36:00 crc kubenswrapper[4807]: I1205 12:36:00.879923 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"8c28b308-84a6-423f-8b24-9814aa6e275b","Type":"ContainerDied","Data":"b440b50270ec6b4666fdf3288b1e1d3171804a477605a28866a45e40416679f3"} Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.161648 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.161868 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="122f1236-de9b-447f-b378-7fa22891f74e" containerName="watcher-decision-engine" containerID="cri-o://92e613d2405d5672afb26257ed113131f48d3ae321d90f120bd7d380646134c1" gracePeriod=30 Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.178837 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.348315 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data-custom\") pod \"8c28b308-84a6-423f-8b24-9814aa6e275b\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.348664 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-cert-memcached-mtls\") pod \"8c28b308-84a6-423f-8b24-9814aa6e275b\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.348732 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-combined-ca-bundle\") pod \"8c28b308-84a6-423f-8b24-9814aa6e275b\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.348759 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c28b308-84a6-423f-8b24-9814aa6e275b-etc-machine-id\") pod \"8c28b308-84a6-423f-8b24-9814aa6e275b\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.348854 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data\") pod \"8c28b308-84a6-423f-8b24-9814aa6e275b\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.348868 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-scripts\") pod \"8c28b308-84a6-423f-8b24-9814aa6e275b\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.348904 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqprb\" (UniqueName: \"kubernetes.io/projected/8c28b308-84a6-423f-8b24-9814aa6e275b-kube-api-access-jqprb\") pod \"8c28b308-84a6-423f-8b24-9814aa6e275b\" (UID: \"8c28b308-84a6-423f-8b24-9814aa6e275b\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.352466 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c28b308-84a6-423f-8b24-9814aa6e275b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8c28b308-84a6-423f-8b24-9814aa6e275b" (UID: "8c28b308-84a6-423f-8b24-9814aa6e275b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.354809 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8c28b308-84a6-423f-8b24-9814aa6e275b" (UID: "8c28b308-84a6-423f-8b24-9814aa6e275b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.357079 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-scripts" (OuterVolumeSpecName: "scripts") pod "8c28b308-84a6-423f-8b24-9814aa6e275b" (UID: "8c28b308-84a6-423f-8b24-9814aa6e275b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.358640 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c28b308-84a6-423f-8b24-9814aa6e275b-kube-api-access-jqprb" (OuterVolumeSpecName: "kube-api-access-jqprb") pod "8c28b308-84a6-423f-8b24-9814aa6e275b" (UID: "8c28b308-84a6-423f-8b24-9814aa6e275b"). InnerVolumeSpecName "kube-api-access-jqprb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.452332 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c28b308-84a6-423f-8b24-9814aa6e275b" (UID: "8c28b308-84a6-423f-8b24-9814aa6e275b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.457687 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.457725 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqprb\" (UniqueName: \"kubernetes.io/projected/8c28b308-84a6-423f-8b24-9814aa6e275b-kube-api-access-jqprb\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.457737 4807 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.457751 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.457763 4807 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c28b308-84a6-423f-8b24-9814aa6e275b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.499720 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data" (OuterVolumeSpecName: "config-data") pod "8c28b308-84a6-423f-8b24-9814aa6e275b" (UID: "8c28b308-84a6-423f-8b24-9814aa6e275b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.516629 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "8c28b308-84a6-423f-8b24-9814aa6e275b" (UID: "8c28b308-84a6-423f-8b24-9814aa6e275b"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.559908 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.559943 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c28b308-84a6-423f-8b24-9814aa6e275b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.626060 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762062 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-dev\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762134 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-lib-cinder\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762158 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-combined-ca-bundle\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762176 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-lib-modules\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762185 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-dev" (OuterVolumeSpecName: "dev") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762238 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-sys\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762256 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-cinder\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762271 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-brick\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762302 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-scripts\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762329 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-sys" (OuterVolumeSpecName: "sys") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762370 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762345 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-machine-id\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762340 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762395 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762433 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-run" (OuterVolumeSpecName: "run") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762415 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-run\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762472 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762509 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfjb2\" (UniqueName: \"kubernetes.io/projected/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-kube-api-access-nfjb2\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762586 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-nvme\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762617 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-iscsi\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762964 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-cert-memcached-mtls\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763010 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data-custom\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763056 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data\") pod \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\" (UID: \"91cfacd9-7fa0-4e43-8a2e-560c74e1a865\") " Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763719 4807 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-dev\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763738 4807 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763748 4807 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-sys\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763757 4807 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763766 4807 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763774 4807 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.763783 4807 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-run\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762681 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762698 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.762300 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.772679 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.776010 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-scripts" (OuterVolumeSpecName: "scripts") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.791309 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-kube-api-access-nfjb2" (OuterVolumeSpecName: "kube-api-access-nfjb2") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "kube-api-access-nfjb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.826843 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.865683 4807 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.865717 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.865730 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfjb2\" (UniqueName: \"kubernetes.io/projected/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-kube-api-access-nfjb2\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.865741 4807 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.865749 4807 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.865759 4807 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.886662 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.904481 4807 generic.go:334] "Generic (PLEG): container finished" podID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerID="3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b" exitCode=0 Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.904566 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"91cfacd9-7fa0-4e43-8a2e-560c74e1a865","Type":"ContainerDied","Data":"3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b"} Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.904595 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"91cfacd9-7fa0-4e43-8a2e-560c74e1a865","Type":"ContainerDied","Data":"cbd8b358f84bc5f2b29b4f11a85977956616165149eb7ae0954374cc94976364"} Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.904622 4807 scope.go:117] "RemoveContainer" containerID="b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.904701 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.913869 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"8c28b308-84a6-423f-8b24-9814aa6e275b","Type":"ContainerDied","Data":"1edf90b737b16307418426d712e140c068986b673c8e88f04cf2d8c3c1791b60"} Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.913969 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.970281 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:01 crc kubenswrapper[4807]: I1205 12:36:01.971676 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data" (OuterVolumeSpecName: "config-data") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.038397 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "91cfacd9-7fa0-4e43-8a2e-560c74e1a865" (UID: "91cfacd9-7fa0-4e43-8a2e-560c74e1a865"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.052934 4807 scope.go:117] "RemoveContainer" containerID="3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.058156 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.071314 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.071340 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91cfacd9-7fa0-4e43-8a2e-560c74e1a865-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.081683 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095232 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.095642 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="probe" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095658 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="probe" Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.095675 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="probe" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095681 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="probe" Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.095706 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="cinder-backup" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095713 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="cinder-backup" Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.095725 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="cinder-scheduler" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095740 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="cinder-scheduler" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095890 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="cinder-scheduler" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095916 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" containerName="probe" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095926 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="probe" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.095936 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" containerName="cinder-backup" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.096812 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.097662 4807 scope.go:117] "RemoveContainer" containerID="b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2" Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.100470 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2\": container with ID starting with b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2 not found: ID does not exist" containerID="b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.100518 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2"} err="failed to get container status \"b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2\": rpc error: code = NotFound desc = could not find container \"b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2\": container with ID starting with b5653867c774bda49c078070e1447a17235966b2a315dc2290f1c77d5a3dc0d2 not found: ID does not exist" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.100563 4807 scope.go:117] "RemoveContainer" containerID="3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.100616 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scheduler-config-data" Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.100970 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b\": container with ID starting with 3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b not found: ID does not exist" containerID="3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.101003 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b"} err="failed to get container status \"3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b\": rpc error: code = NotFound desc = could not find container \"3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b\": container with ID starting with 3ee271a53bb31ec98f6cfe6031af390021a03e2151954ad21def6446eeec1d4b not found: ID does not exist" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.101022 4807 scope.go:117] "RemoveContainer" containerID="57a9bf66514db280695b8253af4214896cd847bc9afc6b711db29158d5e4c62d" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.113120 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.129317 4807 scope.go:117] "RemoveContainer" containerID="b440b50270ec6b4666fdf3288b1e1d3171804a477605a28866a45e40416679f3" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.172582 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.172644 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.172684 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6npk5\" (UniqueName: \"kubernetes.io/projected/c8480492-182b-4dea-9161-111c4eaf27f8-kube-api-access-6npk5\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.172704 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.172736 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.172755 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8480492-182b-4dea-9161-111c4eaf27f8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.173170 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-scripts\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.235775 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.236113 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:36:02 crc kubenswrapper[4807]: E1205 12:36:02.244006 4807 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c28b308_84a6_423f_8b24_9814aa6e275b.slice\": RecentStats: unable to find data in memory cache]" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.245044 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.280694 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-scripts\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.280777 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.280891 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.280989 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6npk5\" (UniqueName: \"kubernetes.io/projected/c8480492-182b-4dea-9161-111c4eaf27f8-kube-api-access-6npk5\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.281029 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.281101 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.281139 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8480492-182b-4dea-9161-111c4eaf27f8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.281601 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8480492-182b-4dea-9161-111c4eaf27f8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.289847 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.291413 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.291653 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.292198 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-scripts\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.294175 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.314706 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.316335 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.318299 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.321446 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6npk5\" (UniqueName: \"kubernetes.io/projected/c8480492-182b-4dea-9161-111c4eaf27f8-kube-api-access-6npk5\") pod \"cinder-scheduler-0\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.324172 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.327346 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-backup-config-data" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.431377 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485193 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-nvme\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485240 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485262 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485278 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-run\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485306 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-scripts\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485324 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-dev\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485342 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485359 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-lib-modules\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485406 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485455 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485491 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485537 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485579 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data-custom\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485604 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjfd\" (UniqueName: \"kubernetes.io/projected/da7800d2-920a-485a-956a-0a894125c7d2-kube-api-access-2tjfd\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485628 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.485779 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-sys\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.588344 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589037 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589079 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589116 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589155 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data-custom\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589187 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjfd\" (UniqueName: \"kubernetes.io/projected/da7800d2-920a-485a-956a-0a894125c7d2-kube-api-access-2tjfd\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589209 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589230 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-sys\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589280 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589286 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-nvme\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589344 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-nvme\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589385 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589411 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-run\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589432 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589497 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-scripts\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589573 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-dev\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589615 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589655 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-lib-modules\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589893 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-lib-modules\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589945 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589983 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-run\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.590016 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.590427 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-sys\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.589242 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.590863 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-dev\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.590952 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.596739 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-scripts\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.597073 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.597652 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.599045 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.599161 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data-custom\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.613129 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjfd\" (UniqueName: \"kubernetes.io/projected/da7800d2-920a-485a-956a-0a894125c7d2-kube-api-access-2tjfd\") pod \"cinder-backup-0\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.640381 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.842934 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.843259 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-central-agent" containerID="cri-o://82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" gracePeriod=30 Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.843392 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="proxy-httpd" containerID="cri-o://dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" gracePeriod=30 Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.843441 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="sg-core" containerID="cri-o://8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" gracePeriod=30 Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.843482 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-notification-agent" containerID="cri-o://e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" gracePeriod=30 Dec 05 12:36:02 crc kubenswrapper[4807]: I1205 12:36:02.940436 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.093341 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.131643 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.246155 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c28b308-84a6-423f-8b24-9814aa6e275b" path="/var/lib/kubelet/pods/8c28b308-84a6-423f-8b24-9814aa6e275b/volumes" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.246888 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91cfacd9-7fa0-4e43-8a2e-560c74e1a865" path="/var/lib/kubelet/pods/91cfacd9-7fa0-4e43-8a2e-560c74e1a865/volumes" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.781748 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928151 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krgkq\" (UniqueName: \"kubernetes.io/projected/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-kube-api-access-krgkq\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928650 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-scripts\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928683 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-config-data\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928729 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-sg-core-conf-yaml\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928759 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-log-httpd\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928779 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-combined-ca-bundle\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928801 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-run-httpd\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.928828 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-ceilometer-tls-certs\") pod \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\" (UID: \"57a61ee2-5709-4a11-b12f-e1b6a9c641f6\") " Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.930225 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.931812 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.937853 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-scripts" (OuterVolumeSpecName: "scripts") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.955406 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-kube-api-access-krgkq" (OuterVolumeSpecName: "kube-api-access-krgkq") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "kube-api-access-krgkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988594 4807 generic.go:334] "Generic (PLEG): container finished" podID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerID="dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" exitCode=0 Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988626 4807 generic.go:334] "Generic (PLEG): container finished" podID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerID="8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" exitCode=2 Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988634 4807 generic.go:334] "Generic (PLEG): container finished" podID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerID="e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" exitCode=0 Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988640 4807 generic.go:334] "Generic (PLEG): container finished" podID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerID="82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" exitCode=0 Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988679 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerDied","Data":"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12"} Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988707 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerDied","Data":"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f"} Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988718 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerDied","Data":"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d"} Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988726 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerDied","Data":"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4"} Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988735 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"57a61ee2-5709-4a11-b12f-e1b6a9c641f6","Type":"ContainerDied","Data":"eb657ddc78fe52bdcaed024debac3dff4153a3c1f713eac39992c6b743ac7c2c"} Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988751 4807 scope.go:117] "RemoveContainer" containerID="dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.988877 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.997883 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"da7800d2-920a-485a-956a-0a894125c7d2","Type":"ContainerStarted","Data":"f2c712447607bcde8dec4a0e187db876ae65229dc0ab8931689c818cc9824779"} Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.997933 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"da7800d2-920a-485a-956a-0a894125c7d2","Type":"ContainerStarted","Data":"697294dc83327a7276e4b71aada5c80eea9aa5a7393e917dfa31edc7a8611c98"} Dec 05 12:36:03 crc kubenswrapper[4807]: I1205 12:36:03.997944 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"da7800d2-920a-485a-956a-0a894125c7d2","Type":"ContainerStarted","Data":"4bf81448736e52b3879f66e7f928ad4d2383cefac275c53e0c33c7a5953b8571"} Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.012677 4807 generic.go:334] "Generic (PLEG): container finished" podID="122f1236-de9b-447f-b378-7fa22891f74e" containerID="92e613d2405d5672afb26257ed113131f48d3ae321d90f120bd7d380646134c1" exitCode=0 Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.012782 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"122f1236-de9b-447f-b378-7fa22891f74e","Type":"ContainerDied","Data":"92e613d2405d5672afb26257ed113131f48d3ae321d90f120bd7d380646134c1"} Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.017864 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"c8480492-182b-4dea-9161-111c4eaf27f8","Type":"ContainerStarted","Data":"826698f3c4780f36488da45c59ac73dbf18ac7af3071b43727f4463416699c1a"} Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.017916 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"c8480492-182b-4dea-9161-111c4eaf27f8","Type":"ContainerStarted","Data":"850f0d9038ae3ab21a1e87461df550bc431df482827ad137c33bb6d020f8aaa9"} Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.032125 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.032156 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.032165 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.032174 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krgkq\" (UniqueName: \"kubernetes.io/projected/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-kube-api-access-krgkq\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.040917 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.048159 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.050967 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-backup-0" podStartSLOduration=2.050930241 podStartE2EDuration="2.050930241s" podCreationTimestamp="2025-12-05 12:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:04.028989235 +0000 UTC m=+1793.522852514" watchObservedRunningTime="2025-12-05 12:36:04.050930241 +0000 UTC m=+1793.544793510" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.133636 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.133667 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.140617 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.173124 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-config-data" (OuterVolumeSpecName: "config-data") pod "57a61ee2-5709-4a11-b12f-e1b6a9c641f6" (UID: "57a61ee2-5709-4a11-b12f-e1b6a9c641f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.217853 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.219717 4807 scope.go:117] "RemoveContainer" containerID="8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.245421 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.245457 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a61ee2-5709-4a11-b12f-e1b6a9c641f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.278196 4807 scope.go:117] "RemoveContainer" containerID="e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.309722 4807 scope.go:117] "RemoveContainer" containerID="82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.343846 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.349164 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-config-data\") pod \"122f1236-de9b-447f-b378-7fa22891f74e\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.349244 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-custom-prometheus-ca\") pod \"122f1236-de9b-447f-b378-7fa22891f74e\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.349273 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-cert-memcached-mtls\") pod \"122f1236-de9b-447f-b378-7fa22891f74e\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.349295 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-combined-ca-bundle\") pod \"122f1236-de9b-447f-b378-7fa22891f74e\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.349335 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89t2q\" (UniqueName: \"kubernetes.io/projected/122f1236-de9b-447f-b378-7fa22891f74e-kube-api-access-89t2q\") pod \"122f1236-de9b-447f-b378-7fa22891f74e\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.349383 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122f1236-de9b-447f-b378-7fa22891f74e-logs\") pod \"122f1236-de9b-447f-b378-7fa22891f74e\" (UID: \"122f1236-de9b-447f-b378-7fa22891f74e\") " Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.355698 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/122f1236-de9b-447f-b378-7fa22891f74e-logs" (OuterVolumeSpecName: "logs") pod "122f1236-de9b-447f-b378-7fa22891f74e" (UID: "122f1236-de9b-447f-b378-7fa22891f74e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.379865 4807 scope.go:117] "RemoveContainer" containerID="dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.381469 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.381956 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": container with ID starting with dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12 not found: ID does not exist" containerID="dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.381994 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12"} err="failed to get container status \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": rpc error: code = NotFound desc = could not find container \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": container with ID starting with dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.382019 4807 scope.go:117] "RemoveContainer" containerID="8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.382235 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": container with ID starting with 8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f not found: ID does not exist" containerID="8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.382261 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f"} err="failed to get container status \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": rpc error: code = NotFound desc = could not find container \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": container with ID starting with 8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.382279 4807 scope.go:117] "RemoveContainer" containerID="e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.382628 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": container with ID starting with e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d not found: ID does not exist" containerID="e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.382656 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d"} err="failed to get container status \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": rpc error: code = NotFound desc = could not find container \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": container with ID starting with e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.382674 4807 scope.go:117] "RemoveContainer" containerID="82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.394417 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_122f1236-de9b-447f-b378-7fa22891f74e/watcher-decision-engine/0.log" Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.404859 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": container with ID starting with 82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4 not found: ID does not exist" containerID="82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.405895 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4"} err="failed to get container status \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": rpc error: code = NotFound desc = could not find container \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": container with ID starting with 82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.405925 4807 scope.go:117] "RemoveContainer" containerID="dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.404894 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122f1236-de9b-447f-b378-7fa22891f74e-kube-api-access-89t2q" (OuterVolumeSpecName: "kube-api-access-89t2q") pod "122f1236-de9b-447f-b378-7fa22891f74e" (UID: "122f1236-de9b-447f-b378-7fa22891f74e"). InnerVolumeSpecName "kube-api-access-89t2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.409712 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12"} err="failed to get container status \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": rpc error: code = NotFound desc = could not find container \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": container with ID starting with dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.409762 4807 scope.go:117] "RemoveContainer" containerID="8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.416707 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f"} err="failed to get container status \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": rpc error: code = NotFound desc = could not find container \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": container with ID starting with 8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.416769 4807 scope.go:117] "RemoveContainer" containerID="e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.419776 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.420172 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-notification-agent" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420195 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-notification-agent" Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.420210 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122f1236-de9b-447f-b378-7fa22891f74e" containerName="watcher-decision-engine" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420217 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="122f1236-de9b-447f-b378-7fa22891f74e" containerName="watcher-decision-engine" Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.420228 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="sg-core" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420234 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="sg-core" Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.420246 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-central-agent" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420252 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-central-agent" Dec 05 12:36:04 crc kubenswrapper[4807]: E1205 12:36:04.420262 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="proxy-httpd" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420268 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="proxy-httpd" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420437 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="122f1236-de9b-447f-b378-7fa22891f74e" containerName="watcher-decision-engine" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420458 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="proxy-httpd" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420466 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="sg-core" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420473 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-central-agent" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420483 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" containerName="ceilometer-notification-agent" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420622 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d"} err="failed to get container status \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": rpc error: code = NotFound desc = could not find container \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": container with ID starting with e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.420646 4807 scope.go:117] "RemoveContainer" containerID="82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.421033 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4"} err="failed to get container status \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": rpc error: code = NotFound desc = could not find container \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": container with ID starting with 82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.421086 4807 scope.go:117] "RemoveContainer" containerID="dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.421657 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12"} err="failed to get container status \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": rpc error: code = NotFound desc = could not find container \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": container with ID starting with dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.421708 4807 scope.go:117] "RemoveContainer" containerID="8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.422056 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.422901 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f"} err="failed to get container status \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": rpc error: code = NotFound desc = could not find container \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": container with ID starting with 8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.422939 4807 scope.go:117] "RemoveContainer" containerID="e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.424781 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d"} err="failed to get container status \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": rpc error: code = NotFound desc = could not find container \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": container with ID starting with e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.424815 4807 scope.go:117] "RemoveContainer" containerID="82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.425578 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4"} err="failed to get container status \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": rpc error: code = NotFound desc = could not find container \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": container with ID starting with 82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.425600 4807 scope.go:117] "RemoveContainer" containerID="dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.425833 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.425978 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.426109 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.436758 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-config-data" (OuterVolumeSpecName: "config-data") pod "122f1236-de9b-447f-b378-7fa22891f74e" (UID: "122f1236-de9b-447f-b378-7fa22891f74e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.443290 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12"} err="failed to get container status \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": rpc error: code = NotFound desc = could not find container \"dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12\": container with ID starting with dced77d374bcf10d4bcaec9c5522fc7e5d7a44a44f0f5cab07b2f5cf9ca69d12 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.443332 4807 scope.go:117] "RemoveContainer" containerID="8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.444242 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f"} err="failed to get container status \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": rpc error: code = NotFound desc = could not find container \"8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f\": container with ID starting with 8058dcf91cd384f56d2244da22701566b763378ccde6cd0f918edebff208280f not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.444282 4807 scope.go:117] "RemoveContainer" containerID="e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.445735 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d"} err="failed to get container status \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": rpc error: code = NotFound desc = could not find container \"e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d\": container with ID starting with e62042fbed5408a887b9239ac21a079e9f8be6073c0fdd6e4549a3a8fdb7f35d not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.445763 4807 scope.go:117] "RemoveContainer" containerID="82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.446294 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4"} err="failed to get container status \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": rpc error: code = NotFound desc = could not find container \"82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4\": container with ID starting with 82373dccb9ffe739b1ea2bbbf1b871da0aa4aaa8fc56cc751e0f1fdf2bb168d4 not found: ID does not exist" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.451737 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.451777 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89t2q\" (UniqueName: \"kubernetes.io/projected/122f1236-de9b-447f-b378-7fa22891f74e-kube-api-access-89t2q\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.451790 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122f1236-de9b-447f-b378-7fa22891f74e-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.451825 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.487751 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "122f1236-de9b-447f-b378-7fa22891f74e" (UID: "122f1236-de9b-447f-b378-7fa22891f74e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.492629 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "122f1236-de9b-447f-b378-7fa22891f74e" (UID: "122f1236-de9b-447f-b378-7fa22891f74e"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.518823 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "122f1236-de9b-447f-b378-7fa22891f74e" (UID: "122f1236-de9b-447f-b378-7fa22891f74e"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555174 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-config-data\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555237 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555297 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-scripts\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555327 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555355 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvjzl\" (UniqueName: \"kubernetes.io/projected/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-kube-api-access-wvjzl\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555406 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555431 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-log-httpd\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555467 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-run-httpd\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555605 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555618 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.555626 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/122f1236-de9b-447f-b378-7fa22891f74e-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.657708 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.657792 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-scripts\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.660929 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.661453 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvjzl\" (UniqueName: \"kubernetes.io/projected/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-kube-api-access-wvjzl\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.661564 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.661601 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-log-httpd\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.662216 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-log-httpd\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.669050 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.672746 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.674907 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-scripts\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.675588 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.675891 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-run-httpd\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.676224 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-config-data\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.676391 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-run-httpd\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.688215 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvjzl\" (UniqueName: \"kubernetes.io/projected/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-kube-api-access-wvjzl\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.690678 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-config-data\") pod \"ceilometer-0\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:04 crc kubenswrapper[4807]: I1205 12:36:04.756107 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.049660 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.057612 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"122f1236-de9b-447f-b378-7fa22891f74e","Type":"ContainerDied","Data":"a047df03065aa05f203c0e71eb6f097e3b6a6d1ff233840ca8e3a7eef8c94493"} Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.057666 4807 scope.go:117] "RemoveContainer" containerID="92e613d2405d5672afb26257ed113131f48d3ae321d90f120bd7d380646134c1" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.153601 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.173825 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.191687 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.192871 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.195366 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.224300 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.233984 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.251177 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="122f1236-de9b-447f-b378-7fa22891f74e" path="/var/lib/kubelet/pods/122f1236-de9b-447f-b378-7fa22891f74e/volumes" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.252325 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a61ee2-5709-4a11-b12f-e1b6a9c641f6" path="/var/lib/kubelet/pods/57a61ee2-5709-4a11-b12f-e1b6a9c641f6/volumes" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.288467 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.288565 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6gs2\" (UniqueName: \"kubernetes.io/projected/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-kube-api-access-l6gs2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.288586 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.288789 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.288856 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.288951 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.390637 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6gs2\" (UniqueName: \"kubernetes.io/projected/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-kube-api-access-l6gs2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.390716 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.390773 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.390801 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.390841 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.390927 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.392089 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.397056 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.397114 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.398666 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.407538 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.419079 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6gs2\" (UniqueName: \"kubernetes.io/projected/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-kube-api-access-l6gs2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:05 crc kubenswrapper[4807]: I1205 12:36:05.556826 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:06 crc kubenswrapper[4807]: I1205 12:36:06.079826 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerStarted","Data":"89941c9eecb433bae1ee49ee24a49672927bca32cfcb648fc1471c2f797f0747"} Dec 05 12:36:06 crc kubenswrapper[4807]: I1205 12:36:06.084391 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"c8480492-182b-4dea-9161-111c4eaf27f8","Type":"ContainerStarted","Data":"3727d4d809af1fa021fe3832c21487a05a49cd4bde677e48cb7b44460fd10992"} Dec 05 12:36:06 crc kubenswrapper[4807]: I1205 12:36:06.095432 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:06 crc kubenswrapper[4807]: I1205 12:36:06.871368 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:36:06 crc kubenswrapper[4807]: I1205 12:36:06.899896 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-scheduler-0" podStartSLOduration=4.899879976 podStartE2EDuration="4.899879976s" podCreationTimestamp="2025-12-05 12:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:06.117146106 +0000 UTC m=+1795.611009365" watchObservedRunningTime="2025-12-05 12:36:06.899879976 +0000 UTC m=+1796.393743245" Dec 05 12:36:07 crc kubenswrapper[4807]: I1205 12:36:07.094602 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ee6685f5-47ba-4dc5-ad06-9805b4c8c898","Type":"ContainerStarted","Data":"57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0"} Dec 05 12:36:07 crc kubenswrapper[4807]: I1205 12:36:07.094963 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ee6685f5-47ba-4dc5-ad06-9805b4c8c898","Type":"ContainerStarted","Data":"f536cca633781425ccac5ac3517024b6b22229dab0ebebd12b857bb7ac0263d7"} Dec 05 12:36:07 crc kubenswrapper[4807]: I1205 12:36:07.099172 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerStarted","Data":"bb9e7ca064f3694851fd091812108d7ea4b587b418e2af0716c82f64282b0714"} Dec 05 12:36:07 crc kubenswrapper[4807]: I1205 12:36:07.099215 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerStarted","Data":"163df5e5085daf43a7689a516521cf66aa1952b4e5bcd8c887e9a6ea48b20237"} Dec 05 12:36:07 crc kubenswrapper[4807]: I1205 12:36:07.120283 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.120263018 podStartE2EDuration="2.120263018s" podCreationTimestamp="2025-12-05 12:36:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:07.117446969 +0000 UTC m=+1796.611310258" watchObservedRunningTime="2025-12-05 12:36:07.120263018 +0000 UTC m=+1796.614126287" Dec 05 12:36:07 crc kubenswrapper[4807]: I1205 12:36:07.432545 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:07 crc kubenswrapper[4807]: I1205 12:36:07.640458 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:08 crc kubenswrapper[4807]: I1205 12:36:08.110757 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerStarted","Data":"6fb5f5be2ad6e7d72458fbfa43f5ef0ec808edd54e2957d8eacddc079db6b65c"} Dec 05 12:36:08 crc kubenswrapper[4807]: I1205 12:36:08.214805 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:09 crc kubenswrapper[4807]: I1205 12:36:09.421768 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:10 crc kubenswrapper[4807]: I1205 12:36:10.141105 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerStarted","Data":"66c31d2ef16ef1dd781137340a0370151465d553a5b91561d0bebddbe9f9e01f"} Dec 05 12:36:10 crc kubenswrapper[4807]: I1205 12:36:10.142403 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:10 crc kubenswrapper[4807]: I1205 12:36:10.166961 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.354531903 podStartE2EDuration="6.166939863s" podCreationTimestamp="2025-12-05 12:36:04 +0000 UTC" firstStartedPulling="2025-12-05 12:36:05.207701878 +0000 UTC m=+1794.701565137" lastFinishedPulling="2025-12-05 12:36:09.020109828 +0000 UTC m=+1798.513973097" observedRunningTime="2025-12-05 12:36:10.162546676 +0000 UTC m=+1799.656409945" watchObservedRunningTime="2025-12-05 12:36:10.166939863 +0000 UTC m=+1799.660803132" Dec 05 12:36:10 crc kubenswrapper[4807]: I1205 12:36:10.627144 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:11 crc kubenswrapper[4807]: I1205 12:36:11.883882 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:12 crc kubenswrapper[4807]: I1205 12:36:12.649211 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:12 crc kubenswrapper[4807]: I1205 12:36:12.995209 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:13 crc kubenswrapper[4807]: I1205 12:36:13.129166 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:14 crc kubenswrapper[4807]: I1205 12:36:14.235330 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:36:14 crc kubenswrapper[4807]: E1205 12:36:14.235918 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:36:14 crc kubenswrapper[4807]: I1205 12:36:14.363481 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:15 crc kubenswrapper[4807]: I1205 12:36:15.557427 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:15 crc kubenswrapper[4807]: I1205 12:36:15.582263 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:15 crc kubenswrapper[4807]: I1205 12:36:15.621975 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:16 crc kubenswrapper[4807]: I1205 12:36:16.194385 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:16 crc kubenswrapper[4807]: I1205 12:36:16.217132 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:16 crc kubenswrapper[4807]: I1205 12:36:16.893842 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.249549 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-p5qc9"] Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.254119 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-p5qc9"] Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.262727 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.273147 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.273399 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="cinder-scheduler" containerID="cri-o://826698f3c4780f36488da45c59ac73dbf18ac7af3071b43727f4463416699c1a" gracePeriod=30 Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.273857 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="probe" containerID="cri-o://3727d4d809af1fa021fe3832c21487a05a49cd4bde677e48cb7b44460fd10992" gracePeriod=30 Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.294816 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.295164 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="cinder-backup" containerID="cri-o://697294dc83327a7276e4b71aada5c80eea9aa5a7393e917dfa31edc7a8611c98" gracePeriod=30 Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.295315 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="probe" containerID="cri-o://f2c712447607bcde8dec4a0e187db876ae65229dc0ab8931689c818cc9824779" gracePeriod=30 Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.370196 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.370681 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api-log" containerID="cri-o://8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131" gracePeriod=30 Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.371144 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api" containerID="cri-o://a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0" gracePeriod=30 Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.406007 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder828f-account-delete-6mbnn"] Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.408651 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.422494 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder828f-account-delete-6mbnn"] Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.501573 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed204a9-1074-4035-a575-b98370727a78-operator-scripts\") pod \"cinder828f-account-delete-6mbnn\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.501759 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvqk\" (UniqueName: \"kubernetes.io/projected/fed204a9-1074-4035-a575-b98370727a78-kube-api-access-lnvqk\") pod \"cinder828f-account-delete-6mbnn\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.603488 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed204a9-1074-4035-a575-b98370727a78-operator-scripts\") pod \"cinder828f-account-delete-6mbnn\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.603660 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvqk\" (UniqueName: \"kubernetes.io/projected/fed204a9-1074-4035-a575-b98370727a78-kube-api-access-lnvqk\") pod \"cinder828f-account-delete-6mbnn\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.604715 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed204a9-1074-4035-a575-b98370727a78-operator-scripts\") pod \"cinder828f-account-delete-6mbnn\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.627257 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvqk\" (UniqueName: \"kubernetes.io/projected/fed204a9-1074-4035-a575-b98370727a78-kube-api-access-lnvqk\") pod \"cinder828f-account-delete-6mbnn\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:17 crc kubenswrapper[4807]: I1205 12:36:17.739306 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:18 crc kubenswrapper[4807]: I1205 12:36:18.217057 4807 generic.go:334] "Generic (PLEG): container finished" podID="63b5b13b-b94f-4208-acda-836de99625f0" containerID="8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131" exitCode=143 Dec 05 12:36:18 crc kubenswrapper[4807]: I1205 12:36:18.217186 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"63b5b13b-b94f-4208-acda-836de99625f0","Type":"ContainerDied","Data":"8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131"} Dec 05 12:36:18 crc kubenswrapper[4807]: I1205 12:36:18.220105 4807 generic.go:334] "Generic (PLEG): container finished" podID="c8480492-182b-4dea-9161-111c4eaf27f8" containerID="3727d4d809af1fa021fe3832c21487a05a49cd4bde677e48cb7b44460fd10992" exitCode=0 Dec 05 12:36:18 crc kubenswrapper[4807]: I1205 12:36:18.220187 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"c8480492-182b-4dea-9161-111c4eaf27f8","Type":"ContainerDied","Data":"3727d4d809af1fa021fe3832c21487a05a49cd4bde677e48cb7b44460fd10992"} Dec 05 12:36:18 crc kubenswrapper[4807]: I1205 12:36:18.253375 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder828f-account-delete-6mbnn"] Dec 05 12:36:18 crc kubenswrapper[4807]: W1205 12:36:18.261543 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfed204a9_1074_4035_a575_b98370727a78.slice/crio-a0ea2a55ea8c21944b71eae107ed74add3d965f9802581548bef46b7921b5219 WatchSource:0}: Error finding container a0ea2a55ea8c21944b71eae107ed74add3d965f9802581548bef46b7921b5219: Status 404 returned error can't find the container with id a0ea2a55ea8c21944b71eae107ed74add3d965f9802581548bef46b7921b5219 Dec 05 12:36:18 crc kubenswrapper[4807]: I1205 12:36:18.525333 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.177842 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.231078 4807 generic.go:334] "Generic (PLEG): container finished" podID="fed204a9-1074-4035-a575-b98370727a78" containerID="ff9557c277c332d4bf78a3efd6fb9d0c81a4fac8cccc80a5a8cc997c94538cf6" exitCode=0 Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.231141 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" event={"ID":"fed204a9-1074-4035-a575-b98370727a78","Type":"ContainerDied","Data":"ff9557c277c332d4bf78a3efd6fb9d0c81a4fac8cccc80a5a8cc997c94538cf6"} Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.231229 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" event={"ID":"fed204a9-1074-4035-a575-b98370727a78","Type":"ContainerStarted","Data":"a0ea2a55ea8c21944b71eae107ed74add3d965f9802581548bef46b7921b5219"} Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.233787 4807 generic.go:334] "Generic (PLEG): container finished" podID="da7800d2-920a-485a-956a-0a894125c7d2" containerID="f2c712447607bcde8dec4a0e187db876ae65229dc0ab8931689c818cc9824779" exitCode=0 Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.233878 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"da7800d2-920a-485a-956a-0a894125c7d2","Type":"ContainerDied","Data":"f2c712447607bcde8dec4a0e187db876ae65229dc0ab8931689c818cc9824779"} Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.234050 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="ee6685f5-47ba-4dc5-ad06-9805b4c8c898" containerName="watcher-decision-engine" containerID="cri-o://57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0" gracePeriod=30 Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.245516 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03760c91-5719-41e3-a9d7-b23301e92be4" path="/var/lib/kubelet/pods/03760c91-5719-41e3-a9d7-b23301e92be4/volumes" Dec 05 12:36:19 crc kubenswrapper[4807]: I1205 12:36:19.811408 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.146027 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.146339 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-central-agent" containerID="cri-o://163df5e5085daf43a7689a516521cf66aa1952b4e5bcd8c887e9a6ea48b20237" gracePeriod=30 Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.146437 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="sg-core" containerID="cri-o://6fb5f5be2ad6e7d72458fbfa43f5ef0ec808edd54e2957d8eacddc079db6b65c" gracePeriod=30 Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.146453 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="proxy-httpd" containerID="cri-o://66c31d2ef16ef1dd781137340a0370151465d553a5b91561d0bebddbe9f9e01f" gracePeriod=30 Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.146453 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-notification-agent" containerID="cri-o://bb9e7ca064f3694851fd091812108d7ea4b587b418e2af0716c82f64282b0714" gracePeriod=30 Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.155860 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.199:3000/\": EOF" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.544944 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/cinder-api-0" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.196:8776/healthcheck\": read tcp 10.217.0.2:39478->10.217.0.196:8776: read: connection reset by peer" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.696828 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.864770 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnvqk\" (UniqueName: \"kubernetes.io/projected/fed204a9-1074-4035-a575-b98370727a78-kube-api-access-lnvqk\") pod \"fed204a9-1074-4035-a575-b98370727a78\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.864841 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed204a9-1074-4035-a575-b98370727a78-operator-scripts\") pod \"fed204a9-1074-4035-a575-b98370727a78\" (UID: \"fed204a9-1074-4035-a575-b98370727a78\") " Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.868756 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fed204a9-1074-4035-a575-b98370727a78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fed204a9-1074-4035-a575-b98370727a78" (UID: "fed204a9-1074-4035-a575-b98370727a78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.883183 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fed204a9-1074-4035-a575-b98370727a78-kube-api-access-lnvqk" (OuterVolumeSpecName: "kube-api-access-lnvqk") pod "fed204a9-1074-4035-a575-b98370727a78" (UID: "fed204a9-1074-4035-a575-b98370727a78"). InnerVolumeSpecName "kube-api-access-lnvqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.972190 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnvqk\" (UniqueName: \"kubernetes.io/projected/fed204a9-1074-4035-a575-b98370727a78-kube-api-access-lnvqk\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.972233 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed204a9-1074-4035-a575-b98370727a78-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:20 crc kubenswrapper[4807]: I1205 12:36:20.996812 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.058768 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175012 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/63b5b13b-b94f-4208-acda-836de99625f0-kube-api-access-kv8z9\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175065 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-scripts\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175131 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data-custom\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175226 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-internal-tls-certs\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175279 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175339 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-combined-ca-bundle\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175373 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63b5b13b-b94f-4208-acda-836de99625f0-logs\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175403 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63b5b13b-b94f-4208-acda-836de99625f0-etc-machine-id\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175423 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-public-tls-certs\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.175460 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-cert-memcached-mtls\") pod \"63b5b13b-b94f-4208-acda-836de99625f0\" (UID: \"63b5b13b-b94f-4208-acda-836de99625f0\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.176004 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63b5b13b-b94f-4208-acda-836de99625f0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.176378 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63b5b13b-b94f-4208-acda-836de99625f0-logs" (OuterVolumeSpecName: "logs") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.180642 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.180765 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b5b13b-b94f-4208-acda-836de99625f0-kube-api-access-kv8z9" (OuterVolumeSpecName: "kube-api-access-kv8z9") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "kube-api-access-kv8z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.181390 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-scripts" (OuterVolumeSpecName: "scripts") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.225986 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data" (OuterVolumeSpecName: "config-data") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.241871 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.269908 4807 generic.go:334] "Generic (PLEG): container finished" podID="63b5b13b-b94f-4208-acda-836de99625f0" containerID="a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0" exitCode=0 Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.270055 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.272603 4807 generic.go:334] "Generic (PLEG): container finished" podID="da7800d2-920a-485a-956a-0a894125c7d2" containerID="697294dc83327a7276e4b71aada5c80eea9aa5a7393e917dfa31edc7a8611c98" exitCode=0 Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.278578 4807 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.278617 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.278626 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.278636 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63b5b13b-b94f-4208-acda-836de99625f0-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.278648 4807 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63b5b13b-b94f-4208-acda-836de99625f0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.278657 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/63b5b13b-b94f-4208-acda-836de99625f0-kube-api-access-kv8z9\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.278671 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.292849 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.293710 4807 generic.go:334] "Generic (PLEG): container finished" podID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerID="66c31d2ef16ef1dd781137340a0370151465d553a5b91561d0bebddbe9f9e01f" exitCode=0 Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.293730 4807 generic.go:334] "Generic (PLEG): container finished" podID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerID="6fb5f5be2ad6e7d72458fbfa43f5ef0ec808edd54e2957d8eacddc079db6b65c" exitCode=2 Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.293738 4807 generic.go:334] "Generic (PLEG): container finished" podID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerID="163df5e5085daf43a7689a516521cf66aa1952b4e5bcd8c887e9a6ea48b20237" exitCode=0 Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.302935 4807 generic.go:334] "Generic (PLEG): container finished" podID="c8480492-182b-4dea-9161-111c4eaf27f8" containerID="826698f3c4780f36488da45c59ac73dbf18ac7af3071b43727f4463416699c1a" exitCode=0 Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.304609 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.319065 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.333808 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "63b5b13b-b94f-4208-acda-836de99625f0" (UID: "63b5b13b-b94f-4208-acda-836de99625f0"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.379865 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.379906 4807 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.379915 4807 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63b5b13b-b94f-4208-acda-836de99625f0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.451937 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"63b5b13b-b94f-4208-acda-836de99625f0","Type":"ContainerDied","Data":"a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.451992 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"63b5b13b-b94f-4208-acda-836de99625f0","Type":"ContainerDied","Data":"a5f7a7c52872218077fffdf36f172f61574142836ff61ac203552c47889d28eb"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452005 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"da7800d2-920a-485a-956a-0a894125c7d2","Type":"ContainerDied","Data":"697294dc83327a7276e4b71aada5c80eea9aa5a7393e917dfa31edc7a8611c98"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452018 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerDied","Data":"66c31d2ef16ef1dd781137340a0370151465d553a5b91561d0bebddbe9f9e01f"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452029 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerDied","Data":"6fb5f5be2ad6e7d72458fbfa43f5ef0ec808edd54e2957d8eacddc079db6b65c"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452038 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerDied","Data":"163df5e5085daf43a7689a516521cf66aa1952b4e5bcd8c887e9a6ea48b20237"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452046 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"c8480492-182b-4dea-9161-111c4eaf27f8","Type":"ContainerDied","Data":"826698f3c4780f36488da45c59ac73dbf18ac7af3071b43727f4463416699c1a"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452074 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder828f-account-delete-6mbnn" event={"ID":"fed204a9-1074-4035-a575-b98370727a78","Type":"ContainerDied","Data":"a0ea2a55ea8c21944b71eae107ed74add3d965f9802581548bef46b7921b5219"} Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452086 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0ea2a55ea8c21944b71eae107ed74add3d965f9802581548bef46b7921b5219" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.452101 4807 scope.go:117] "RemoveContainer" containerID="a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.467421 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.475487 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.497256 4807 scope.go:117] "RemoveContainer" containerID="8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511030 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8480492-182b-4dea-9161-111c4eaf27f8-etc-machine-id\") pod \"c8480492-182b-4dea-9161-111c4eaf27f8\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511099 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-combined-ca-bundle\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511151 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6npk5\" (UniqueName: \"kubernetes.io/projected/c8480492-182b-4dea-9161-111c4eaf27f8-kube-api-access-6npk5\") pod \"c8480492-182b-4dea-9161-111c4eaf27f8\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511213 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data-custom\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511249 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-cert-memcached-mtls\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511275 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-dev\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511300 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-nvme\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511339 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data-custom\") pod \"c8480492-182b-4dea-9161-111c4eaf27f8\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511362 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-machine-id\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511384 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-combined-ca-bundle\") pod \"c8480492-182b-4dea-9161-111c4eaf27f8\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511404 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511434 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-lib-modules\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511478 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-lib-cinder\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511504 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-run\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511558 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-cert-memcached-mtls\") pod \"c8480492-182b-4dea-9161-111c4eaf27f8\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511583 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-iscsi\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511607 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-scripts\") pod \"c8480492-182b-4dea-9161-111c4eaf27f8\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511628 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-cinder\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511729 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tjfd\" (UniqueName: \"kubernetes.io/projected/da7800d2-920a-485a-956a-0a894125c7d2-kube-api-access-2tjfd\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511778 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data\") pod \"c8480492-182b-4dea-9161-111c4eaf27f8\" (UID: \"c8480492-182b-4dea-9161-111c4eaf27f8\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511804 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-sys\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511834 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-scripts\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.511877 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-brick\") pod \"da7800d2-920a-485a-956a-0a894125c7d2\" (UID: \"da7800d2-920a-485a-956a-0a894125c7d2\") " Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.512297 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.512345 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.512370 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.512396 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-run" (OuterVolumeSpecName: "run") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.513562 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.513598 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.514163 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-dev" (OuterVolumeSpecName: "dev") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.515106 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-sys" (OuterVolumeSpecName: "sys") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.517965 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.520424 4807 scope.go:117] "RemoveContainer" containerID="a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.520578 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.524753 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8480492-182b-4dea-9161-111c4eaf27f8-kube-api-access-6npk5" (OuterVolumeSpecName: "kube-api-access-6npk5") pod "c8480492-182b-4dea-9161-111c4eaf27f8" (UID: "c8480492-182b-4dea-9161-111c4eaf27f8"). InnerVolumeSpecName "kube-api-access-6npk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: E1205 12:36:21.526387 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0\": container with ID starting with a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0 not found: ID does not exist" containerID="a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.526428 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0"} err="failed to get container status \"a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0\": rpc error: code = NotFound desc = could not find container \"a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0\": container with ID starting with a8e8aa247b82121777b6b007fb605235bea92f71ad301fb780e314d9520afdb0 not found: ID does not exist" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.526453 4807 scope.go:117] "RemoveContainer" containerID="8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.527390 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8480492-182b-4dea-9161-111c4eaf27f8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c8480492-182b-4dea-9161-111c4eaf27f8" (UID: "c8480492-182b-4dea-9161-111c4eaf27f8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.528325 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c8480492-182b-4dea-9161-111c4eaf27f8" (UID: "c8480492-182b-4dea-9161-111c4eaf27f8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.528655 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7800d2-920a-485a-956a-0a894125c7d2-kube-api-access-2tjfd" (OuterVolumeSpecName: "kube-api-access-2tjfd") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "kube-api-access-2tjfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.528947 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-scripts" (OuterVolumeSpecName: "scripts") pod "c8480492-182b-4dea-9161-111c4eaf27f8" (UID: "c8480492-182b-4dea-9161-111c4eaf27f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: E1205 12:36:21.535092 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131\": container with ID starting with 8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131 not found: ID does not exist" containerID="8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.535144 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131"} err="failed to get container status \"8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131\": rpc error: code = NotFound desc = could not find container \"8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131\": container with ID starting with 8d70d948e947c57f8e9e00a4c7f96bd12f739b181b3a8c60590a6cc702d71131 not found: ID does not exist" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.541915 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-scripts" (OuterVolumeSpecName: "scripts") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.542027 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.598991 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.606611 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8480492-182b-4dea-9161-111c4eaf27f8" (UID: "c8480492-182b-4dea-9161-111c4eaf27f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616005 4807 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-sys\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616144 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616200 4807 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616272 4807 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8480492-182b-4dea-9161-111c4eaf27f8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616324 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616373 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6npk5\" (UniqueName: \"kubernetes.io/projected/c8480492-182b-4dea-9161-111c4eaf27f8-kube-api-access-6npk5\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616428 4807 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616482 4807 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-dev\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616561 4807 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616618 4807 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616669 4807 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616719 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616780 4807 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616832 4807 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616882 4807 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-run\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.616933 4807 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.617109 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.617187 4807 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/da7800d2-920a-485a-956a-0a894125c7d2-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.617296 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tjfd\" (UniqueName: \"kubernetes.io/projected/da7800d2-920a-485a-956a-0a894125c7d2-kube-api-access-2tjfd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.646119 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data" (OuterVolumeSpecName: "config-data") pod "c8480492-182b-4dea-9161-111c4eaf27f8" (UID: "c8480492-182b-4dea-9161-111c4eaf27f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.658656 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data" (OuterVolumeSpecName: "config-data") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.670137 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.676742 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.685352 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "da7800d2-920a-485a-956a-0a894125c7d2" (UID: "da7800d2-920a-485a-956a-0a894125c7d2"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.718501 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "c8480492-182b-4dea-9161-111c4eaf27f8" (UID: "c8480492-182b-4dea-9161-111c4eaf27f8"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.719840 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.719883 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7800d2-920a-485a-956a-0a894125c7d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.719896 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:21 crc kubenswrapper[4807]: I1205 12:36:21.719907 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8480492-182b-4dea-9161-111c4eaf27f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.305206 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.315050 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"c8480492-182b-4dea-9161-111c4eaf27f8","Type":"ContainerDied","Data":"850f0d9038ae3ab21a1e87461df550bc431df482827ad137c33bb6d020f8aaa9"} Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.315692 4807 scope.go:117] "RemoveContainer" containerID="3727d4d809af1fa021fe3832c21487a05a49cd4bde677e48cb7b44460fd10992" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.315926 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.324769 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"da7800d2-920a-485a-956a-0a894125c7d2","Type":"ContainerDied","Data":"4bf81448736e52b3879f66e7f928ad4d2383cefac275c53e0c33c7a5953b8571"} Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.324867 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.357096 4807 scope.go:117] "RemoveContainer" containerID="826698f3c4780f36488da45c59ac73dbf18ac7af3071b43727f4463416699c1a" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.368591 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.377242 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-db-create-rmhlb"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.380461 4807 scope.go:117] "RemoveContainer" containerID="f2c712447607bcde8dec4a0e187db876ae65229dc0ab8931689c818cc9824779" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.387165 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.395278 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-db-create-rmhlb"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.401751 4807 scope.go:117] "RemoveContainer" containerID="697294dc83327a7276e4b71aada5c80eea9aa5a7393e917dfa31edc7a8611c98" Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.412159 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.418995 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder828f-account-delete-6mbnn"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.427072 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.434729 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-828f-account-create-update-bwwwm"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.441499 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder828f-account-delete-6mbnn"] Dec 05 12:36:22 crc kubenswrapper[4807]: I1205 12:36:22.447078 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-828f-account-create-update-bwwwm"] Dec 05 12:36:23 crc kubenswrapper[4807]: I1205 12:36:23.284448 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63b5b13b-b94f-4208-acda-836de99625f0" path="/var/lib/kubelet/pods/63b5b13b-b94f-4208-acda-836de99625f0/volumes" Dec 05 12:36:23 crc kubenswrapper[4807]: I1205 12:36:23.285548 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" path="/var/lib/kubelet/pods/c8480492-182b-4dea-9161-111c4eaf27f8/volumes" Dec 05 12:36:23 crc kubenswrapper[4807]: I1205 12:36:23.286116 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2cd9d02-3e04-48f9-9c0d-21d36772ac95" path="/var/lib/kubelet/pods/d2cd9d02-3e04-48f9-9c0d-21d36772ac95/volumes" Dec 05 12:36:23 crc kubenswrapper[4807]: I1205 12:36:23.287109 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da7800d2-920a-485a-956a-0a894125c7d2" path="/var/lib/kubelet/pods/da7800d2-920a-485a-956a-0a894125c7d2/volumes" Dec 05 12:36:23 crc kubenswrapper[4807]: I1205 12:36:23.287697 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe7aa02-168c-490f-88d0-57de6b64f790" path="/var/lib/kubelet/pods/efe7aa02-168c-490f-88d0-57de6b64f790/volumes" Dec 05 12:36:23 crc kubenswrapper[4807]: I1205 12:36:23.288228 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fed204a9-1074-4035-a575-b98370727a78" path="/var/lib/kubelet/pods/fed204a9-1074-4035-a575-b98370727a78/volumes" Dec 05 12:36:23 crc kubenswrapper[4807]: I1205 12:36:23.529347 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:24 crc kubenswrapper[4807]: I1205 12:36:24.767858 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:25 crc kubenswrapper[4807]: I1205 12:36:25.235314 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:36:25 crc kubenswrapper[4807]: E1205 12:36:25.235555 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:36:26 crc kubenswrapper[4807]: I1205 12:36:26.029232 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.289643 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.424471 4807 generic.go:334] "Generic (PLEG): container finished" podID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerID="bb9e7ca064f3694851fd091812108d7ea4b587b418e2af0716c82f64282b0714" exitCode=0 Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.424549 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerDied","Data":"bb9e7ca064f3694851fd091812108d7ea4b587b418e2af0716c82f64282b0714"} Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.714071 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.838964 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-config-data\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839229 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-combined-ca-bundle\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839272 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-run-httpd\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839335 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-log-httpd\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839355 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvjzl\" (UniqueName: \"kubernetes.io/projected/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-kube-api-access-wvjzl\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839459 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-ceilometer-tls-certs\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839496 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-scripts\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839570 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-sg-core-conf-yaml\") pod \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\" (UID: \"d64ef835-2c2b-4b79-aa1e-b636b503ffa2\") " Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.839638 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.840003 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.840234 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.847117 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-scripts" (OuterVolumeSpecName: "scripts") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.847865 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-kube-api-access-wvjzl" (OuterVolumeSpecName: "kube-api-access-wvjzl") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "kube-api-access-wvjzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.867656 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.894159 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.915138 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.931458 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-config-data" (OuterVolumeSpecName: "config-data") pod "d64ef835-2c2b-4b79-aa1e-b636b503ffa2" (UID: "d64ef835-2c2b-4b79-aa1e-b636b503ffa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.941200 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.941229 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.941238 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.941246 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.941255 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvjzl\" (UniqueName: \"kubernetes.io/projected/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-kube-api-access-wvjzl\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.941267 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4807]: I1205 12:36:27.941276 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d64ef835-2c2b-4b79-aa1e-b636b503ffa2-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.159056 4807 scope.go:117] "RemoveContainer" containerID="27f5a60abfaa0dc45005768ba8aaca84e93fa903456e06e70936227755c71f43" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.178875 4807 scope.go:117] "RemoveContainer" containerID="1e548a724927c630d01a9d517cca27e1992007825cd3f82099c91aa6a04323aa" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.195415 4807 scope.go:117] "RemoveContainer" containerID="714f34656212ba81b5b11ebf4ec2fdd4e513048f8881def632a5dd9892cd0144" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.214017 4807 scope.go:117] "RemoveContainer" containerID="898fbca429efeb2b2180ddce3fa3cedbb54b0d9ed16e5f37f99f1da44cc99e29" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.233428 4807 scope.go:117] "RemoveContainer" containerID="891b042e34e110dbafab55b31116053a5b0da80b0c9c167fc0ee570ba70e50d9" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.260919 4807 scope.go:117] "RemoveContainer" containerID="0f8a686a5d0c75dd0634f89ca363f77558309b1add3d617015bdb9227a8e6b75" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.280445 4807 scope.go:117] "RemoveContainer" containerID="703fb818c19be18bd95560b9e803aceee8efb580fec44fe5714bfff738e5773d" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.465314 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"d64ef835-2c2b-4b79-aa1e-b636b503ffa2","Type":"ContainerDied","Data":"89941c9eecb433bae1ee49ee24a49672927bca32cfcb648fc1471c2f797f0747"} Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.465358 4807 scope.go:117] "RemoveContainer" containerID="66c31d2ef16ef1dd781137340a0370151465d553a5b91561d0bebddbe9f9e01f" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.465479 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.502765 4807 scope.go:117] "RemoveContainer" containerID="6fb5f5be2ad6e7d72458fbfa43f5ef0ec808edd54e2957d8eacddc079db6b65c" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.513521 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_ee6685f5-47ba-4dc5-ad06-9805b4c8c898/watcher-decision-engine/0.log" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.533465 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.540252 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.553188 4807 scope.go:117] "RemoveContainer" containerID="bb9e7ca064f3694851fd091812108d7ea4b587b418e2af0716c82f64282b0714" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.558250 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.559292 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="cinder-backup" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.559476 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="cinder-backup" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.559571 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-central-agent" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.559653 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-central-agent" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.559729 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="probe" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.559832 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="probe" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.559916 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-notification-agent" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.559985 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-notification-agent" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.560056 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="cinder-scheduler" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.560520 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="cinder-scheduler" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.560645 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="sg-core" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.560721 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="sg-core" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.560807 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="probe" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.560887 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="probe" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.560968 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed204a9-1074-4035-a575-b98370727a78" containerName="mariadb-account-delete" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.561048 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed204a9-1074-4035-a575-b98370727a78" containerName="mariadb-account-delete" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.561156 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.561240 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.561327 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api-log" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.561403 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api-log" Dec 05 12:36:28 crc kubenswrapper[4807]: E1205 12:36:28.561491 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="proxy-httpd" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.561611 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="proxy-httpd" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.561993 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-central-agent" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.562091 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.562169 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="ceilometer-notification-agent" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.562240 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="proxy-httpd" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.562653 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="probe" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.562761 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b5b13b-b94f-4208-acda-836de99625f0" containerName="cinder-api-log" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.562845 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="probe" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.562930 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" containerName="sg-core" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.563011 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed204a9-1074-4035-a575-b98370727a78" containerName="mariadb-account-delete" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.563094 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7800d2-920a-485a-956a-0a894125c7d2" containerName="cinder-backup" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.563166 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8480492-182b-4dea-9161-111c4eaf27f8" containerName="cinder-scheduler" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.565151 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.568402 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.568680 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.568862 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.572274 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.604798 4807 scope.go:117] "RemoveContainer" containerID="163df5e5085daf43a7689a516521cf66aa1952b4e5bcd8c887e9a6ea48b20237" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.656453 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-run-httpd\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.656638 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.656764 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzpns\" (UniqueName: \"kubernetes.io/projected/0519017c-161f-41df-8641-632b6778a4e2-kube-api-access-jzpns\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.656870 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.656951 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-config-data\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.657045 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-scripts\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.657149 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-log-httpd\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.657223 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759323 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-log-httpd\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759378 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759431 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-run-httpd\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759451 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759491 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzpns\" (UniqueName: \"kubernetes.io/projected/0519017c-161f-41df-8641-632b6778a4e2-kube-api-access-jzpns\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759575 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759600 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-config-data\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.759630 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-scripts\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.760073 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-log-httpd\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.760562 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-run-httpd\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.773077 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-scripts\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.773075 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.773317 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.773327 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-config-data\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.774172 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.777224 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzpns\" (UniqueName: \"kubernetes.io/projected/0519017c-161f-41df-8641-632b6778a4e2-kube-api-access-jzpns\") pod \"ceilometer-0\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:28 crc kubenswrapper[4807]: I1205 12:36:28.922410 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.245713 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64ef835-2c2b-4b79-aa1e-b636b503ffa2" path="/var/lib/kubelet/pods/d64ef835-2c2b-4b79-aa1e-b636b503ffa2/volumes" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.272752 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.371741 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6gs2\" (UniqueName: \"kubernetes.io/projected/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-kube-api-access-l6gs2\") pod \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.371807 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-cert-memcached-mtls\") pod \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.371875 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-logs\") pod \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.371942 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-config-data\") pod \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.372025 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-combined-ca-bundle\") pod \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.372070 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-custom-prometheus-ca\") pod \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\" (UID: \"ee6685f5-47ba-4dc5-ad06-9805b4c8c898\") " Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.372338 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-logs" (OuterVolumeSpecName: "logs") pod "ee6685f5-47ba-4dc5-ad06-9805b4c8c898" (UID: "ee6685f5-47ba-4dc5-ad06-9805b4c8c898"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.373104 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.379666 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-kube-api-access-l6gs2" (OuterVolumeSpecName: "kube-api-access-l6gs2") pod "ee6685f5-47ba-4dc5-ad06-9805b4c8c898" (UID: "ee6685f5-47ba-4dc5-ad06-9805b4c8c898"). InnerVolumeSpecName "kube-api-access-l6gs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.394110 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee6685f5-47ba-4dc5-ad06-9805b4c8c898" (UID: "ee6685f5-47ba-4dc5-ad06-9805b4c8c898"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.395258 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "ee6685f5-47ba-4dc5-ad06-9805b4c8c898" (UID: "ee6685f5-47ba-4dc5-ad06-9805b4c8c898"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.448428 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-config-data" (OuterVolumeSpecName: "config-data") pod "ee6685f5-47ba-4dc5-ad06-9805b4c8c898" (UID: "ee6685f5-47ba-4dc5-ad06-9805b4c8c898"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.450431 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "ee6685f5-47ba-4dc5-ad06-9805b4c8c898" (UID: "ee6685f5-47ba-4dc5-ad06-9805b4c8c898"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.474564 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.474605 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.474614 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6gs2\" (UniqueName: \"kubernetes.io/projected/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-kube-api-access-l6gs2\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.474626 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.474635 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6685f5-47ba-4dc5-ad06-9805b4c8c898-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.481834 4807 generic.go:334] "Generic (PLEG): container finished" podID="ee6685f5-47ba-4dc5-ad06-9805b4c8c898" containerID="57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0" exitCode=0 Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.481920 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ee6685f5-47ba-4dc5-ad06-9805b4c8c898","Type":"ContainerDied","Data":"57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0"} Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.481956 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"ee6685f5-47ba-4dc5-ad06-9805b4c8c898","Type":"ContainerDied","Data":"f536cca633781425ccac5ac3517024b6b22229dab0ebebd12b857bb7ac0263d7"} Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.481981 4807 scope.go:117] "RemoveContainer" containerID="57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.482083 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.482569 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.505915 4807 scope.go:117] "RemoveContainer" containerID="57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0" Dec 05 12:36:29 crc kubenswrapper[4807]: E1205 12:36:29.508956 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0\": container with ID starting with 57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0 not found: ID does not exist" containerID="57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.508995 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0"} err="failed to get container status \"57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0\": rpc error: code = NotFound desc = could not find container \"57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0\": container with ID starting with 57f29b34d84564bca63992a776f687d445d91a291800f9bec37df006997ddfb0 not found: ID does not exist" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.510962 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.517895 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.528668 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:29 crc kubenswrapper[4807]: E1205 12:36:29.529121 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee6685f5-47ba-4dc5-ad06-9805b4c8c898" containerName="watcher-decision-engine" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.529144 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6685f5-47ba-4dc5-ad06-9805b4c8c898" containerName="watcher-decision-engine" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.529365 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee6685f5-47ba-4dc5-ad06-9805b4c8c898" containerName="watcher-decision-engine" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.530159 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.535888 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.543756 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.677953 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.678314 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.678354 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wmll\" (UniqueName: \"kubernetes.io/projected/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-kube-api-access-8wmll\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.678370 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.678390 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.678762 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.779757 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.780078 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.780192 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.780283 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wmll\" (UniqueName: \"kubernetes.io/projected/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-kube-api-access-8wmll\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.780387 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.780463 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.780890 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.789166 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.789256 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.789292 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.789484 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.795730 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wmll\" (UniqueName: \"kubernetes.io/projected/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-kube-api-access-8wmll\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:29 crc kubenswrapper[4807]: I1205 12:36:29.864902 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:30 crc kubenswrapper[4807]: W1205 12:36:30.293336 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod057ae85f_e82d_49a6_a96b_4b0d450b1b3c.slice/crio-dea0dcc580c9c795f62627f28556241efa50969179d26c86ba2cb0e35c84fd34 WatchSource:0}: Error finding container dea0dcc580c9c795f62627f28556241efa50969179d26c86ba2cb0e35c84fd34: Status 404 returned error can't find the container with id dea0dcc580c9c795f62627f28556241efa50969179d26c86ba2cb0e35c84fd34 Dec 05 12:36:30 crc kubenswrapper[4807]: I1205 12:36:30.294407 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:30 crc kubenswrapper[4807]: I1205 12:36:30.494446 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"057ae85f-e82d-49a6-a96b-4b0d450b1b3c","Type":"ContainerStarted","Data":"d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805"} Dec 05 12:36:30 crc kubenswrapper[4807]: I1205 12:36:30.494759 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"057ae85f-e82d-49a6-a96b-4b0d450b1b3c","Type":"ContainerStarted","Data":"dea0dcc580c9c795f62627f28556241efa50969179d26c86ba2cb0e35c84fd34"} Dec 05 12:36:30 crc kubenswrapper[4807]: I1205 12:36:30.496475 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerStarted","Data":"edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a"} Dec 05 12:36:30 crc kubenswrapper[4807]: I1205 12:36:30.496505 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerStarted","Data":"193984debab6ea8d27b1d2c22d0b9d36cdbba06afd00989723862b4f27e2900b"} Dec 05 12:36:30 crc kubenswrapper[4807]: I1205 12:36:30.979180 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:31 crc kubenswrapper[4807]: I1205 12:36:31.254541 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee6685f5-47ba-4dc5-ad06-9805b4c8c898" path="/var/lib/kubelet/pods/ee6685f5-47ba-4dc5-ad06-9805b4c8c898/volumes" Dec 05 12:36:31 crc kubenswrapper[4807]: I1205 12:36:31.265791 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.265771692 podStartE2EDuration="2.265771692s" podCreationTimestamp="2025-12-05 12:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:30.514702435 +0000 UTC m=+1820.008565714" watchObservedRunningTime="2025-12-05 12:36:31.265771692 +0000 UTC m=+1820.759634961" Dec 05 12:36:31 crc kubenswrapper[4807]: I1205 12:36:31.516554 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerStarted","Data":"a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa"} Dec 05 12:36:31 crc kubenswrapper[4807]: I1205 12:36:31.516605 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerStarted","Data":"278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531"} Dec 05 12:36:32 crc kubenswrapper[4807]: I1205 12:36:32.173201 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:33 crc kubenswrapper[4807]: I1205 12:36:33.450274 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:33 crc kubenswrapper[4807]: I1205 12:36:33.541181 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerStarted","Data":"b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e"} Dec 05 12:36:33 crc kubenswrapper[4807]: I1205 12:36:33.541317 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:33 crc kubenswrapper[4807]: I1205 12:36:33.567170 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.557925758 podStartE2EDuration="5.56714992s" podCreationTimestamp="2025-12-05 12:36:28 +0000 UTC" firstStartedPulling="2025-12-05 12:36:29.490287993 +0000 UTC m=+1818.984151262" lastFinishedPulling="2025-12-05 12:36:32.499512155 +0000 UTC m=+1821.993375424" observedRunningTime="2025-12-05 12:36:33.560711673 +0000 UTC m=+1823.054574962" watchObservedRunningTime="2025-12-05 12:36:33.56714992 +0000 UTC m=+1823.061013189" Dec 05 12:36:34 crc kubenswrapper[4807]: I1205 12:36:34.642799 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:35 crc kubenswrapper[4807]: I1205 12:36:35.888623 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:36 crc kubenswrapper[4807]: I1205 12:36:36.236343 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:36:36 crc kubenswrapper[4807]: E1205 12:36:36.236717 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:36:37 crc kubenswrapper[4807]: I1205 12:36:37.108419 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:38 crc kubenswrapper[4807]: I1205 12:36:38.382609 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:39 crc kubenswrapper[4807]: I1205 12:36:39.581626 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:39 crc kubenswrapper[4807]: I1205 12:36:39.866751 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:39 crc kubenswrapper[4807]: I1205 12:36:39.905873 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:40 crc kubenswrapper[4807]: I1205 12:36:40.596027 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:40 crc kubenswrapper[4807]: I1205 12:36:40.626805 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:40 crc kubenswrapper[4807]: I1205 12:36:40.850027 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.108746 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_057ae85f-e82d-49a6-a96b-4b0d450b1b3c/watcher-decision-engine/0.log" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.242505 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw"] Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.250974 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-nsvdw"] Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.300143 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher4cf0-account-delete-mbzhw"] Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.301226 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.321712 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher4cf0-account-delete-mbzhw"] Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.328060 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.378126 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8wxj\" (UniqueName: \"kubernetes.io/projected/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-kube-api-access-g8wxj\") pod \"watcher4cf0-account-delete-mbzhw\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.378210 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-operator-scripts\") pod \"watcher4cf0-account-delete-mbzhw\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.411818 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.417259 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-kuttl-api-log" containerID="cri-o://20a268d732dca89e82a6f4b36161c1aac4100e84574aaef120d01f5e3b2139dc" gracePeriod=30 Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.417287 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-api" containerID="cri-o://79dac9f110987b8941c35b7544a549ef83c6afe470d621093d7551ff45659185" gracePeriod=30 Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.451839 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.452432 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="6735e09d-799c-4aed-ac7a-02e301da235f" containerName="watcher-applier" containerID="cri-o://da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32" gracePeriod=30 Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.495843 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-operator-scripts\") pod \"watcher4cf0-account-delete-mbzhw\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.495975 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8wxj\" (UniqueName: \"kubernetes.io/projected/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-kube-api-access-g8wxj\") pod \"watcher4cf0-account-delete-mbzhw\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.496857 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-operator-scripts\") pod \"watcher4cf0-account-delete-mbzhw\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.524407 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8wxj\" (UniqueName: \"kubernetes.io/projected/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-kube-api-access-g8wxj\") pod \"watcher4cf0-account-delete-mbzhw\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.611250 4807 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" secret="" err="secret \"watcher-watcher-kuttl-dockercfg-xvs4w\" not found" Dec 05 12:36:42 crc kubenswrapper[4807]: I1205 12:36:42.622627 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:42 crc kubenswrapper[4807]: E1205 12:36:42.700800 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:42 crc kubenswrapper[4807]: E1205 12:36:42.700889 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data podName:057ae85f-e82d-49a6-a96b-4b0d450b1b3c nodeName:}" failed. No retries permitted until 2025-12-05 12:36:43.200868371 +0000 UTC m=+1832.694731640 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c") : secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.156501 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher4cf0-account-delete-mbzhw"] Dec 05 12:36:43 crc kubenswrapper[4807]: E1205 12:36:43.209926 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:43 crc kubenswrapper[4807]: E1205 12:36:43.210222 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data podName:057ae85f-e82d-49a6-a96b-4b0d450b1b3c nodeName:}" failed. No retries permitted until 2025-12-05 12:36:44.210204717 +0000 UTC m=+1833.704067986 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c") : secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.252040 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73daef9b-354f-4347-bbcd-32b8375a579e" path="/var/lib/kubelet/pods/73daef9b-354f-4347-bbcd-32b8375a579e/volumes" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.632514 4807 generic.go:334] "Generic (PLEG): container finished" podID="c4778f9b-c90d-4db1-aa33-f4bc601ae82b" containerID="7a936439c18f8b9d66bd1647bb08616ea470c1fee1b313845a5a94efeb102f53" exitCode=0 Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.632642 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" event={"ID":"c4778f9b-c90d-4db1-aa33-f4bc601ae82b","Type":"ContainerDied","Data":"7a936439c18f8b9d66bd1647bb08616ea470c1fee1b313845a5a94efeb102f53"} Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.632682 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" event={"ID":"c4778f9b-c90d-4db1-aa33-f4bc601ae82b","Type":"ContainerStarted","Data":"f63eed11c5298a9450ed584c23dcc72c22f4ee4313ee78afe459380705a951bd"} Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.636162 4807 generic.go:334] "Generic (PLEG): container finished" podID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerID="79dac9f110987b8941c35b7544a549ef83c6afe470d621093d7551ff45659185" exitCode=0 Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.636204 4807 generic.go:334] "Generic (PLEG): container finished" podID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerID="20a268d732dca89e82a6f4b36161c1aac4100e84574aaef120d01f5e3b2139dc" exitCode=143 Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.636335 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"410711c8-f736-4f77-ae89-e43fe721e7cd","Type":"ContainerDied","Data":"79dac9f110987b8941c35b7544a549ef83c6afe470d621093d7551ff45659185"} Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.636386 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"410711c8-f736-4f77-ae89-e43fe721e7cd","Type":"ContainerDied","Data":"20a268d732dca89e82a6f4b36161c1aac4100e84574aaef120d01f5e3b2139dc"} Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.636406 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="057ae85f-e82d-49a6-a96b-4b0d450b1b3c" containerName="watcher-decision-engine" containerID="cri-o://d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805" gracePeriod=30 Dec 05 12:36:43 crc kubenswrapper[4807]: E1205 12:36:43.820401 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:36:43 crc kubenswrapper[4807]: E1205 12:36:43.821879 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:36:43 crc kubenswrapper[4807]: E1205 12:36:43.823560 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:36:43 crc kubenswrapper[4807]: E1205 12:36:43.823605 4807 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="6735e09d-799c-4aed-ac7a-02e301da235f" containerName="watcher-applier" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.875349 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.924996 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-combined-ca-bundle\") pod \"410711c8-f736-4f77-ae89-e43fe721e7cd\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.925047 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-custom-prometheus-ca\") pod \"410711c8-f736-4f77-ae89-e43fe721e7cd\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.925097 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-cert-memcached-mtls\") pod \"410711c8-f736-4f77-ae89-e43fe721e7cd\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.925125 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-config-data\") pod \"410711c8-f736-4f77-ae89-e43fe721e7cd\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.925153 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410711c8-f736-4f77-ae89-e43fe721e7cd-logs\") pod \"410711c8-f736-4f77-ae89-e43fe721e7cd\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.925173 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4sb7\" (UniqueName: \"kubernetes.io/projected/410711c8-f736-4f77-ae89-e43fe721e7cd-kube-api-access-v4sb7\") pod \"410711c8-f736-4f77-ae89-e43fe721e7cd\" (UID: \"410711c8-f736-4f77-ae89-e43fe721e7cd\") " Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.926171 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410711c8-f736-4f77-ae89-e43fe721e7cd-logs" (OuterVolumeSpecName: "logs") pod "410711c8-f736-4f77-ae89-e43fe721e7cd" (UID: "410711c8-f736-4f77-ae89-e43fe721e7cd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.931076 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410711c8-f736-4f77-ae89-e43fe721e7cd-kube-api-access-v4sb7" (OuterVolumeSpecName: "kube-api-access-v4sb7") pod "410711c8-f736-4f77-ae89-e43fe721e7cd" (UID: "410711c8-f736-4f77-ae89-e43fe721e7cd"). InnerVolumeSpecName "kube-api-access-v4sb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.960516 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "410711c8-f736-4f77-ae89-e43fe721e7cd" (UID: "410711c8-f736-4f77-ae89-e43fe721e7cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.963346 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "410711c8-f736-4f77-ae89-e43fe721e7cd" (UID: "410711c8-f736-4f77-ae89-e43fe721e7cd"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.980070 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-config-data" (OuterVolumeSpecName: "config-data") pod "410711c8-f736-4f77-ae89-e43fe721e7cd" (UID: "410711c8-f736-4f77-ae89-e43fe721e7cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:43 crc kubenswrapper[4807]: I1205 12:36:43.999410 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "410711c8-f736-4f77-ae89-e43fe721e7cd" (UID: "410711c8-f736-4f77-ae89-e43fe721e7cd"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.026611 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.026644 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.026657 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.026669 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410711c8-f736-4f77-ae89-e43fe721e7cd-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.026680 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410711c8-f736-4f77-ae89-e43fe721e7cd-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.026689 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4sb7\" (UniqueName: \"kubernetes.io/projected/410711c8-f736-4f77-ae89-e43fe721e7cd-kube-api-access-v4sb7\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:44 crc kubenswrapper[4807]: E1205 12:36:44.230646 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:44 crc kubenswrapper[4807]: E1205 12:36:44.230749 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data podName:057ae85f-e82d-49a6-a96b-4b0d450b1b3c nodeName:}" failed. No retries permitted until 2025-12-05 12:36:46.230726913 +0000 UTC m=+1835.724590182 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c") : secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.648158 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"410711c8-f736-4f77-ae89-e43fe721e7cd","Type":"ContainerDied","Data":"7e189443e35a0db158d2568a741481f5be9af51df14e22a063801db86e42a47e"} Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.648226 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.648486 4807 scope.go:117] "RemoveContainer" containerID="79dac9f110987b8941c35b7544a549ef83c6afe470d621093d7551ff45659185" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.684263 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.685196 4807 scope.go:117] "RemoveContainer" containerID="20a268d732dca89e82a6f4b36161c1aac4100e84574aaef120d01f5e3b2139dc" Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.692002 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.825456 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.825753 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-central-agent" containerID="cri-o://edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a" gracePeriod=30 Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.825818 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="proxy-httpd" containerID="cri-o://b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e" gracePeriod=30 Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.825871 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="sg-core" containerID="cri-o://a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa" gracePeriod=30 Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.825901 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-notification-agent" containerID="cri-o://278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531" gracePeriod=30 Dec 05 12:36:44 crc kubenswrapper[4807]: I1205 12:36:44.847916 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.202:3000/\": EOF" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.040469 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.145196 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-operator-scripts\") pod \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.145272 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8wxj\" (UniqueName: \"kubernetes.io/projected/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-kube-api-access-g8wxj\") pod \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\" (UID: \"c4778f9b-c90d-4db1-aa33-f4bc601ae82b\") " Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.146197 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c4778f9b-c90d-4db1-aa33-f4bc601ae82b" (UID: "c4778f9b-c90d-4db1-aa33-f4bc601ae82b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.150708 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-kube-api-access-g8wxj" (OuterVolumeSpecName: "kube-api-access-g8wxj") pod "c4778f9b-c90d-4db1-aa33-f4bc601ae82b" (UID: "c4778f9b-c90d-4db1-aa33-f4bc601ae82b"). InnerVolumeSpecName "kube-api-access-g8wxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.246612 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.246641 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8wxj\" (UniqueName: \"kubernetes.io/projected/c4778f9b-c90d-4db1-aa33-f4bc601ae82b-kube-api-access-g8wxj\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.253948 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" path="/var/lib/kubelet/pods/410711c8-f736-4f77-ae89-e43fe721e7cd/volumes" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.659031 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" event={"ID":"c4778f9b-c90d-4db1-aa33-f4bc601ae82b","Type":"ContainerDied","Data":"f63eed11c5298a9450ed584c23dcc72c22f4ee4313ee78afe459380705a951bd"} Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.659082 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f63eed11c5298a9450ed584c23dcc72c22f4ee4313ee78afe459380705a951bd" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.659042 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher4cf0-account-delete-mbzhw" Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.662586 4807 generic.go:334] "Generic (PLEG): container finished" podID="0519017c-161f-41df-8641-632b6778a4e2" containerID="b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e" exitCode=0 Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.662621 4807 generic.go:334] "Generic (PLEG): container finished" podID="0519017c-161f-41df-8641-632b6778a4e2" containerID="a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa" exitCode=2 Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.662635 4807 generic.go:334] "Generic (PLEG): container finished" podID="0519017c-161f-41df-8641-632b6778a4e2" containerID="edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a" exitCode=0 Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.662654 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerDied","Data":"b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e"} Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.662705 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerDied","Data":"a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa"} Dec 05 12:36:45 crc kubenswrapper[4807]: I1205 12:36:45.662721 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerDied","Data":"edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a"} Dec 05 12:36:46 crc kubenswrapper[4807]: E1205 12:36:46.265095 4807 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:46 crc kubenswrapper[4807]: E1205 12:36:46.265518 4807 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data podName:057ae85f-e82d-49a6-a96b-4b0d450b1b3c nodeName:}" failed. No retries permitted until 2025-12-05 12:36:50.265495932 +0000 UTC m=+1839.759359201 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c") : secret "watcher-kuttl-decision-engine-config-data" not found Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.285024 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.345625 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-n25l8"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.355792 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-n25l8"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.380084 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-scripts\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.380124 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-ceilometer-tls-certs\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.380156 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-config-data\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.380218 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-log-httpd\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.380317 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher4cf0-account-delete-mbzhw"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.382360 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-sg-core-conf-yaml\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.385291 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-run-httpd\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.385380 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzpns\" (UniqueName: \"kubernetes.io/projected/0519017c-161f-41df-8641-632b6778a4e2-kube-api-access-jzpns\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.385426 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-combined-ca-bundle\") pod \"0519017c-161f-41df-8641-632b6778a4e2\" (UID: \"0519017c-161f-41df-8641-632b6778a4e2\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.386483 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.386886 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.421320 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-scripts" (OuterVolumeSpecName: "scripts") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.434871 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0519017c-161f-41df-8641-632b6778a4e2-kube-api-access-jzpns" (OuterVolumeSpecName: "kube-api-access-jzpns") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "kube-api-access-jzpns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.459704 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.470438 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher4cf0-account-delete-mbzhw"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.478447 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-4cf0-account-create-update-6jjf2"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.488817 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.488844 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.488854 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0519017c-161f-41df-8641-632b6778a4e2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.488863 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzpns\" (UniqueName: \"kubernetes.io/projected/0519017c-161f-41df-8641-632b6778a4e2-kube-api-access-jzpns\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.491916 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.492981 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.511789 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.564827 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.589872 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-logs\") pod \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.589920 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wmll\" (UniqueName: \"kubernetes.io/projected/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-kube-api-access-8wmll\") pod \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.589976 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-cert-memcached-mtls\") pod \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.589996 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-custom-prometheus-ca\") pod \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.590054 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data\") pod \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.590083 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-combined-ca-bundle\") pod \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\" (UID: \"057ae85f-e82d-49a6-a96b-4b0d450b1b3c\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.590273 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-logs" (OuterVolumeSpecName: "logs") pod "057ae85f-e82d-49a6-a96b-4b0d450b1b3c" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.590530 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.590560 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.590571 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.590580 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.593115 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-config-data" (OuterVolumeSpecName: "config-data") pod "0519017c-161f-41df-8641-632b6778a4e2" (UID: "0519017c-161f-41df-8641-632b6778a4e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.594869 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-kube-api-access-8wmll" (OuterVolumeSpecName: "kube-api-access-8wmll") pod "057ae85f-e82d-49a6-a96b-4b0d450b1b3c" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c"). InnerVolumeSpecName "kube-api-access-8wmll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.619899 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "057ae85f-e82d-49a6-a96b-4b0d450b1b3c" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.623811 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "057ae85f-e82d-49a6-a96b-4b0d450b1b3c" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.644029 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data" (OuterVolumeSpecName: "config-data") pod "057ae85f-e82d-49a6-a96b-4b0d450b1b3c" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.681517 4807 generic.go:334] "Generic (PLEG): container finished" podID="6735e09d-799c-4aed-ac7a-02e301da235f" containerID="da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32" exitCode=0 Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.681851 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6735e09d-799c-4aed-ac7a-02e301da235f","Type":"ContainerDied","Data":"da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32"} Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.683219 4807 generic.go:334] "Generic (PLEG): container finished" podID="057ae85f-e82d-49a6-a96b-4b0d450b1b3c" containerID="d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805" exitCode=0 Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.683376 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"057ae85f-e82d-49a6-a96b-4b0d450b1b3c","Type":"ContainerDied","Data":"d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805"} Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.683487 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"057ae85f-e82d-49a6-a96b-4b0d450b1b3c","Type":"ContainerDied","Data":"dea0dcc580c9c795f62627f28556241efa50969179d26c86ba2cb0e35c84fd34"} Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.683575 4807 scope.go:117] "RemoveContainer" containerID="d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.683737 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.687813 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "057ae85f-e82d-49a6-a96b-4b0d450b1b3c" (UID: "057ae85f-e82d-49a6-a96b-4b0d450b1b3c"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.688770 4807 generic.go:334] "Generic (PLEG): container finished" podID="0519017c-161f-41df-8641-632b6778a4e2" containerID="278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531" exitCode=0 Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.688860 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerDied","Data":"278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531"} Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.688892 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0519017c-161f-41df-8641-632b6778a4e2","Type":"ContainerDied","Data":"193984debab6ea8d27b1d2c22d0b9d36cdbba06afd00989723862b4f27e2900b"} Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.688993 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.694960 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0519017c-161f-41df-8641-632b6778a4e2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.694994 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wmll\" (UniqueName: \"kubernetes.io/projected/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-kube-api-access-8wmll\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.695007 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.695018 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.695028 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.695038 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057ae85f-e82d-49a6-a96b-4b0d450b1b3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.743853 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.749847 4807 scope.go:117] "RemoveContainer" containerID="d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.750677 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805\": container with ID starting with d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805 not found: ID does not exist" containerID="d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.750900 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805"} err="failed to get container status \"d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805\": rpc error: code = NotFound desc = could not find container \"d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805\": container with ID starting with d3fb96a1b383e95a49f5514023dd936b1c208db955919076fb4c398b97cc4805 not found: ID does not exist" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.750998 4807 scope.go:117] "RemoveContainer" containerID="b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.779942 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788160 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788609 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4778f9b-c90d-4db1-aa33-f4bc601ae82b" containerName="mariadb-account-delete" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788623 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4778f9b-c90d-4db1-aa33-f4bc601ae82b" containerName="mariadb-account-delete" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788638 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="proxy-httpd" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788645 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="proxy-httpd" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788654 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057ae85f-e82d-49a6-a96b-4b0d450b1b3c" containerName="watcher-decision-engine" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788661 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="057ae85f-e82d-49a6-a96b-4b0d450b1b3c" containerName="watcher-decision-engine" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788672 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-api" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788677 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-api" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788692 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-central-agent" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788698 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-central-agent" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788709 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-kuttl-api-log" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788715 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-kuttl-api-log" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788721 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-notification-agent" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788726 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-notification-agent" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.788739 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="sg-core" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788746 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="sg-core" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788914 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-api" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788924 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-kuttl-api-log" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788933 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="sg-core" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788943 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="proxy-httpd" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788958 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4778f9b-c90d-4db1-aa33-f4bc601ae82b" containerName="mariadb-account-delete" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788971 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="057ae85f-e82d-49a6-a96b-4b0d450b1b3c" containerName="watcher-decision-engine" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788978 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-notification-agent" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.788994 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="0519017c-161f-41df-8641-632b6778a4e2" containerName="ceilometer-central-agent" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.791553 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.797918 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.799062 4807 scope.go:117] "RemoveContainer" containerID="a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.801384 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.801653 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.802268 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.830948 4807 scope.go:117] "RemoveContainer" containerID="278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.870746 4807 scope.go:117] "RemoveContainer" containerID="edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897655 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897710 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-log-httpd\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897744 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897778 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-scripts\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897808 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-config-data\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897829 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897864 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-run-httpd\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.897897 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2lc8\" (UniqueName: \"kubernetes.io/projected/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-kube-api-access-l2lc8\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.900707 4807 scope.go:117] "RemoveContainer" containerID="b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.901111 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e\": container with ID starting with b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e not found: ID does not exist" containerID="b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.901148 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e"} err="failed to get container status \"b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e\": rpc error: code = NotFound desc = could not find container \"b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e\": container with ID starting with b45c4fc2d77b2162efaca708e32c9374a375f3ae09c2067a059fea8c081bc54e not found: ID does not exist" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.901173 4807 scope.go:117] "RemoveContainer" containerID="a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.901658 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa\": container with ID starting with a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa not found: ID does not exist" containerID="a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.901685 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa"} err="failed to get container status \"a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa\": rpc error: code = NotFound desc = could not find container \"a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa\": container with ID starting with a7fc2ef2a23a2b36aa8ac92827079f8aa416e2266d02871dc915aac7427691fa not found: ID does not exist" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.901701 4807 scope.go:117] "RemoveContainer" containerID="278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.901925 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531\": container with ID starting with 278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531 not found: ID does not exist" containerID="278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.901955 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531"} err="failed to get container status \"278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531\": rpc error: code = NotFound desc = could not find container \"278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531\": container with ID starting with 278253cc6b171432ec434bb79e749eae8da457b2a83c4dccb914d17561da6531 not found: ID does not exist" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.901973 4807 scope.go:117] "RemoveContainer" containerID="edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a" Dec 05 12:36:47 crc kubenswrapper[4807]: E1205 12:36:47.902298 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a\": container with ID starting with edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a not found: ID does not exist" containerID="edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.902325 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a"} err="failed to get container status \"edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a\": rpc error: code = NotFound desc = could not find container \"edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a\": container with ID starting with edd03a04bb74e9304fd8b6ac18cdd57555bdbeda1c7b7a9898a5d53446ce559a not found: ID does not exist" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.962757 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.998641 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-combined-ca-bundle\") pod \"6735e09d-799c-4aed-ac7a-02e301da235f\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.998702 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6735e09d-799c-4aed-ac7a-02e301da235f-logs\") pod \"6735e09d-799c-4aed-ac7a-02e301da235f\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " Dec 05 12:36:47 crc kubenswrapper[4807]: I1205 12:36:47.998746 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv2rj\" (UniqueName: \"kubernetes.io/projected/6735e09d-799c-4aed-ac7a-02e301da235f-kube-api-access-pv2rj\") pod \"6735e09d-799c-4aed-ac7a-02e301da235f\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.998932 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-cert-memcached-mtls\") pod \"6735e09d-799c-4aed-ac7a-02e301da235f\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999008 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-config-data\") pod \"6735e09d-799c-4aed-ac7a-02e301da235f\" (UID: \"6735e09d-799c-4aed-ac7a-02e301da235f\") " Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999219 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999243 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-log-httpd\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999271 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999305 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-scripts\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999334 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-config-data\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999360 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999396 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-run-httpd\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999434 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2lc8\" (UniqueName: \"kubernetes.io/projected/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-kube-api-access-l2lc8\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:47.999588 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6735e09d-799c-4aed-ac7a-02e301da235f-logs" (OuterVolumeSpecName: "logs") pod "6735e09d-799c-4aed-ac7a-02e301da235f" (UID: "6735e09d-799c-4aed-ac7a-02e301da235f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.000074 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-run-httpd\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.000082 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-log-httpd\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.003013 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6735e09d-799c-4aed-ac7a-02e301da235f-kube-api-access-pv2rj" (OuterVolumeSpecName: "kube-api-access-pv2rj") pod "6735e09d-799c-4aed-ac7a-02e301da235f" (UID: "6735e09d-799c-4aed-ac7a-02e301da235f"). InnerVolumeSpecName "kube-api-access-pv2rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.009029 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-scripts\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.009520 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-config-data\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.033575 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.042066 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2lc8\" (UniqueName: \"kubernetes.io/projected/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-kube-api-access-l2lc8\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.043661 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.048839 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.060363 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6735e09d-799c-4aed-ac7a-02e301da235f" (UID: "6735e09d-799c-4aed-ac7a-02e301da235f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.082231 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-config-data" (OuterVolumeSpecName: "config-data") pod "6735e09d-799c-4aed-ac7a-02e301da235f" (UID: "6735e09d-799c-4aed-ac7a-02e301da235f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.101294 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.101328 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.101340 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6735e09d-799c-4aed-ac7a-02e301da235f-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.101348 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv2rj\" (UniqueName: \"kubernetes.io/projected/6735e09d-799c-4aed-ac7a-02e301da235f-kube-api-access-pv2rj\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.109800 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.110563 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "6735e09d-799c-4aed-ac7a-02e301da235f" (UID: "6735e09d-799c-4aed-ac7a-02e301da235f"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.202597 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6735e09d-799c-4aed-ac7a-02e301da235f-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.206436 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.220891 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.577085 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:36:48 crc kubenswrapper[4807]: W1205 12:36:48.582729 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebda6f78_25bc_4bb4_bedd_d5c7a7285bd4.slice/crio-acf0a0864d53847eb5014f2fec8b2300ead0baad6605d033d6115785dc94b40a WatchSource:0}: Error finding container acf0a0864d53847eb5014f2fec8b2300ead0baad6605d033d6115785dc94b40a: Status 404 returned error can't find the container with id acf0a0864d53847eb5014f2fec8b2300ead0baad6605d033d6115785dc94b40a Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.689339 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.186:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.689335 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="410711c8-f736-4f77-ae89-e43fe721e7cd" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.0.186:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.698213 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6735e09d-799c-4aed-ac7a-02e301da235f","Type":"ContainerDied","Data":"6287e56bc5e87f516b1f41e40b4a59835a721d42e5f7617cb9662f399c5c9cee"} Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.698275 4807 scope.go:117] "RemoveContainer" containerID="da687ff92107d564c237052e23fa50b9cbc7718ea849dc9b76a8c87611e6be32" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.698227 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.703088 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerStarted","Data":"acf0a0864d53847eb5014f2fec8b2300ead0baad6605d033d6115785dc94b40a"} Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.737720 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:36:48 crc kubenswrapper[4807]: I1205 12:36:48.743430 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.260347 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0519017c-161f-41df-8641-632b6778a4e2" path="/var/lib/kubelet/pods/0519017c-161f-41df-8641-632b6778a4e2/volumes" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.261794 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057ae85f-e82d-49a6-a96b-4b0d450b1b3c" path="/var/lib/kubelet/pods/057ae85f-e82d-49a6-a96b-4b0d450b1b3c/volumes" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.262412 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a873051-f632-40ff-8086-c1017d8f09c7" path="/var/lib/kubelet/pods/5a873051-f632-40ff-8086-c1017d8f09c7/volumes" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.263663 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6735e09d-799c-4aed-ac7a-02e301da235f" path="/var/lib/kubelet/pods/6735e09d-799c-4aed-ac7a-02e301da235f/volumes" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.264234 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="904acf87-6567-4146-84d2-38c23643b7d6" path="/var/lib/kubelet/pods/904acf87-6567-4146-84d2-38c23643b7d6/volumes" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.264798 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4778f9b-c90d-4db1-aa33-f4bc601ae82b" path="/var/lib/kubelet/pods/c4778f9b-c90d-4db1-aa33-f4bc601ae82b/volumes" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.471132 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-vjbb7"] Dec 05 12:36:49 crc kubenswrapper[4807]: E1205 12:36:49.471469 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6735e09d-799c-4aed-ac7a-02e301da235f" containerName="watcher-applier" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.471486 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="6735e09d-799c-4aed-ac7a-02e301da235f" containerName="watcher-applier" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.471660 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="6735e09d-799c-4aed-ac7a-02e301da235f" containerName="watcher-applier" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.472176 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.482642 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vjbb7"] Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.490703 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q"] Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.495101 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.499624 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.523623 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj7zd\" (UniqueName: \"kubernetes.io/projected/3997f615-34be-4ff6-96ce-bf70ed5c2248-kube-api-access-gj7zd\") pod \"watcher-db-create-vjbb7\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.523692 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f11e9916-7522-454b-9341-9bc202abde3b-operator-scripts\") pod \"watcher-fc03-account-create-update-7fl7q\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.523725 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2zrk\" (UniqueName: \"kubernetes.io/projected/f11e9916-7522-454b-9341-9bc202abde3b-kube-api-access-h2zrk\") pod \"watcher-fc03-account-create-update-7fl7q\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.523753 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3997f615-34be-4ff6-96ce-bf70ed5c2248-operator-scripts\") pod \"watcher-db-create-vjbb7\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.527333 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q"] Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.625421 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj7zd\" (UniqueName: \"kubernetes.io/projected/3997f615-34be-4ff6-96ce-bf70ed5c2248-kube-api-access-gj7zd\") pod \"watcher-db-create-vjbb7\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.625514 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f11e9916-7522-454b-9341-9bc202abde3b-operator-scripts\") pod \"watcher-fc03-account-create-update-7fl7q\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.625580 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2zrk\" (UniqueName: \"kubernetes.io/projected/f11e9916-7522-454b-9341-9bc202abde3b-kube-api-access-h2zrk\") pod \"watcher-fc03-account-create-update-7fl7q\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.625633 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3997f615-34be-4ff6-96ce-bf70ed5c2248-operator-scripts\") pod \"watcher-db-create-vjbb7\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.626281 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f11e9916-7522-454b-9341-9bc202abde3b-operator-scripts\") pod \"watcher-fc03-account-create-update-7fl7q\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.626499 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3997f615-34be-4ff6-96ce-bf70ed5c2248-operator-scripts\") pod \"watcher-db-create-vjbb7\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.642066 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj7zd\" (UniqueName: \"kubernetes.io/projected/3997f615-34be-4ff6-96ce-bf70ed5c2248-kube-api-access-gj7zd\") pod \"watcher-db-create-vjbb7\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.642329 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2zrk\" (UniqueName: \"kubernetes.io/projected/f11e9916-7522-454b-9341-9bc202abde3b-kube-api-access-h2zrk\") pod \"watcher-fc03-account-create-update-7fl7q\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.716879 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerStarted","Data":"b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af"} Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.794724 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:49 crc kubenswrapper[4807]: I1205 12:36:49.824275 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.235855 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:36:50 crc kubenswrapper[4807]: E1205 12:36:50.236447 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.303903 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vjbb7"] Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.406195 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q"] Dec 05 12:36:50 crc kubenswrapper[4807]: W1205 12:36:50.418737 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf11e9916_7522_454b_9341_9bc202abde3b.slice/crio-a8b0862dc7f223de8e4c8cd5e10182ea0c4a821496b3a41702950babf2cf172d WatchSource:0}: Error finding container a8b0862dc7f223de8e4c8cd5e10182ea0c4a821496b3a41702950babf2cf172d: Status 404 returned error can't find the container with id a8b0862dc7f223de8e4c8cd5e10182ea0c4a821496b3a41702950babf2cf172d Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.731261 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" event={"ID":"f11e9916-7522-454b-9341-9bc202abde3b","Type":"ContainerStarted","Data":"9fbd2ecb81731c6c95659d8d3b5e7ef166d645a9911815a165f17e00793137c8"} Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.731331 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" event={"ID":"f11e9916-7522-454b-9341-9bc202abde3b","Type":"ContainerStarted","Data":"a8b0862dc7f223de8e4c8cd5e10182ea0c4a821496b3a41702950babf2cf172d"} Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.733238 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-vjbb7" event={"ID":"3997f615-34be-4ff6-96ce-bf70ed5c2248","Type":"ContainerStarted","Data":"a72af4e002c5d14d59d9b27a24133ce103a8c6fdc779ef88799a428132c69fd0"} Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.733306 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-vjbb7" event={"ID":"3997f615-34be-4ff6-96ce-bf70ed5c2248","Type":"ContainerStarted","Data":"0d3d41d4c0ae85c4fcc9b5d3d176879e3dfd087c415b9cab96ed3fdfd9b7681f"} Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.736742 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerStarted","Data":"15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9"} Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.736766 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerStarted","Data":"d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b"} Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.745957 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" podStartSLOduration=1.7459379959999999 podStartE2EDuration="1.745937996s" podCreationTimestamp="2025-12-05 12:36:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:50.745018093 +0000 UTC m=+1840.238881362" watchObservedRunningTime="2025-12-05 12:36:50.745937996 +0000 UTC m=+1840.239801265" Dec 05 12:36:50 crc kubenswrapper[4807]: I1205 12:36:50.767468 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-db-create-vjbb7" podStartSLOduration=1.76744963 podStartE2EDuration="1.76744963s" podCreationTimestamp="2025-12-05 12:36:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:50.762189332 +0000 UTC m=+1840.256052611" watchObservedRunningTime="2025-12-05 12:36:50.76744963 +0000 UTC m=+1840.261312899" Dec 05 12:36:51 crc kubenswrapper[4807]: I1205 12:36:51.455242 4807 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podfed204a9-1074-4035-a575-b98370727a78"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podfed204a9-1074-4035-a575-b98370727a78] : Timed out while waiting for systemd to remove kubepods-besteffort-podfed204a9_1074_4035_a575_b98370727a78.slice" Dec 05 12:36:51 crc kubenswrapper[4807]: I1205 12:36:51.747396 4807 generic.go:334] "Generic (PLEG): container finished" podID="f11e9916-7522-454b-9341-9bc202abde3b" containerID="9fbd2ecb81731c6c95659d8d3b5e7ef166d645a9911815a165f17e00793137c8" exitCode=0 Dec 05 12:36:51 crc kubenswrapper[4807]: I1205 12:36:51.747457 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" event={"ID":"f11e9916-7522-454b-9341-9bc202abde3b","Type":"ContainerDied","Data":"9fbd2ecb81731c6c95659d8d3b5e7ef166d645a9911815a165f17e00793137c8"} Dec 05 12:36:51 crc kubenswrapper[4807]: I1205 12:36:51.749333 4807 generic.go:334] "Generic (PLEG): container finished" podID="3997f615-34be-4ff6-96ce-bf70ed5c2248" containerID="a72af4e002c5d14d59d9b27a24133ce103a8c6fdc779ef88799a428132c69fd0" exitCode=0 Dec 05 12:36:51 crc kubenswrapper[4807]: I1205 12:36:51.749371 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-vjbb7" event={"ID":"3997f615-34be-4ff6-96ce-bf70ed5c2248","Type":"ContainerDied","Data":"a72af4e002c5d14d59d9b27a24133ce103a8c6fdc779ef88799a428132c69fd0"} Dec 05 12:36:52 crc kubenswrapper[4807]: I1205 12:36:52.761475 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerStarted","Data":"5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342"} Dec 05 12:36:52 crc kubenswrapper[4807]: I1205 12:36:52.761705 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:36:52 crc kubenswrapper[4807]: I1205 12:36:52.793995 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.384450359 podStartE2EDuration="5.793977279s" podCreationTimestamp="2025-12-05 12:36:47 +0000 UTC" firstStartedPulling="2025-12-05 12:36:48.584690204 +0000 UTC m=+1838.078553503" lastFinishedPulling="2025-12-05 12:36:51.994217154 +0000 UTC m=+1841.488080423" observedRunningTime="2025-12-05 12:36:52.792681227 +0000 UTC m=+1842.286544496" watchObservedRunningTime="2025-12-05 12:36:52.793977279 +0000 UTC m=+1842.287840548" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.182608 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.187969 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.278741 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj7zd\" (UniqueName: \"kubernetes.io/projected/3997f615-34be-4ff6-96ce-bf70ed5c2248-kube-api-access-gj7zd\") pod \"3997f615-34be-4ff6-96ce-bf70ed5c2248\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.278814 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f11e9916-7522-454b-9341-9bc202abde3b-operator-scripts\") pod \"f11e9916-7522-454b-9341-9bc202abde3b\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.278858 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2zrk\" (UniqueName: \"kubernetes.io/projected/f11e9916-7522-454b-9341-9bc202abde3b-kube-api-access-h2zrk\") pod \"f11e9916-7522-454b-9341-9bc202abde3b\" (UID: \"f11e9916-7522-454b-9341-9bc202abde3b\") " Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.278962 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3997f615-34be-4ff6-96ce-bf70ed5c2248-operator-scripts\") pod \"3997f615-34be-4ff6-96ce-bf70ed5c2248\" (UID: \"3997f615-34be-4ff6-96ce-bf70ed5c2248\") " Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.280365 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f11e9916-7522-454b-9341-9bc202abde3b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f11e9916-7522-454b-9341-9bc202abde3b" (UID: "f11e9916-7522-454b-9341-9bc202abde3b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.280845 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3997f615-34be-4ff6-96ce-bf70ed5c2248-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3997f615-34be-4ff6-96ce-bf70ed5c2248" (UID: "3997f615-34be-4ff6-96ce-bf70ed5c2248"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.287343 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f11e9916-7522-454b-9341-9bc202abde3b-kube-api-access-h2zrk" (OuterVolumeSpecName: "kube-api-access-h2zrk") pod "f11e9916-7522-454b-9341-9bc202abde3b" (UID: "f11e9916-7522-454b-9341-9bc202abde3b"). InnerVolumeSpecName "kube-api-access-h2zrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.289648 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3997f615-34be-4ff6-96ce-bf70ed5c2248-kube-api-access-gj7zd" (OuterVolumeSpecName: "kube-api-access-gj7zd") pod "3997f615-34be-4ff6-96ce-bf70ed5c2248" (UID: "3997f615-34be-4ff6-96ce-bf70ed5c2248"). InnerVolumeSpecName "kube-api-access-gj7zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.380095 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj7zd\" (UniqueName: \"kubernetes.io/projected/3997f615-34be-4ff6-96ce-bf70ed5c2248-kube-api-access-gj7zd\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.380143 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f11e9916-7522-454b-9341-9bc202abde3b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.380156 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2zrk\" (UniqueName: \"kubernetes.io/projected/f11e9916-7522-454b-9341-9bc202abde3b-kube-api-access-h2zrk\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.380169 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3997f615-34be-4ff6-96ce-bf70ed5c2248-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.772051 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-vjbb7" event={"ID":"3997f615-34be-4ff6-96ce-bf70ed5c2248","Type":"ContainerDied","Data":"0d3d41d4c0ae85c4fcc9b5d3d176879e3dfd087c415b9cab96ed3fdfd9b7681f"} Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.772334 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d3d41d4c0ae85c4fcc9b5d3d176879e3dfd087c415b9cab96ed3fdfd9b7681f" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.772108 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vjbb7" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.774302 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.782163 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q" event={"ID":"f11e9916-7522-454b-9341-9bc202abde3b","Type":"ContainerDied","Data":"a8b0862dc7f223de8e4c8cd5e10182ea0c4a821496b3a41702950babf2cf172d"} Dec 05 12:36:53 crc kubenswrapper[4807]: I1205 12:36:53.782201 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8b0862dc7f223de8e4c8cd5e10182ea0c4a821496b3a41702950babf2cf172d" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.839547 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm"] Dec 05 12:36:54 crc kubenswrapper[4807]: E1205 12:36:54.839920 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11e9916-7522-454b-9341-9bc202abde3b" containerName="mariadb-account-create-update" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.839938 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11e9916-7522-454b-9341-9bc202abde3b" containerName="mariadb-account-create-update" Dec 05 12:36:54 crc kubenswrapper[4807]: E1205 12:36:54.839969 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3997f615-34be-4ff6-96ce-bf70ed5c2248" containerName="mariadb-database-create" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.839977 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="3997f615-34be-4ff6-96ce-bf70ed5c2248" containerName="mariadb-database-create" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.840175 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f11e9916-7522-454b-9341-9bc202abde3b" containerName="mariadb-account-create-update" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.840196 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="3997f615-34be-4ff6-96ce-bf70ed5c2248" containerName="mariadb-database-create" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.840713 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.843147 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-z96pj" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.843190 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.860746 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm"] Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.899703 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.899752 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.899839 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqxxj\" (UniqueName: \"kubernetes.io/projected/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-kube-api-access-wqxxj\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:54 crc kubenswrapper[4807]: I1205 12:36:54.899969 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-config-data\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.001095 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.001227 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.001270 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqxxj\" (UniqueName: \"kubernetes.io/projected/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-kube-api-access-wqxxj\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.001343 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-config-data\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.006739 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.007374 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-config-data\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.008297 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.029156 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqxxj\" (UniqueName: \"kubernetes.io/projected/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-kube-api-access-wqxxj\") pod \"watcher-kuttl-db-sync-vj6rm\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.160164 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.661086 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm"] Dec 05 12:36:55 crc kubenswrapper[4807]: I1205 12:36:55.790070 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" event={"ID":"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2","Type":"ContainerStarted","Data":"f1e49a4da551e7e94fe7f55f31fb6f3a59346f3651a98accce879a61e812eb6b"} Dec 05 12:36:56 crc kubenswrapper[4807]: I1205 12:36:56.803682 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" event={"ID":"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2","Type":"ContainerStarted","Data":"dc3ff126ea0ecf2857481fe9694669ba2cd1d87a737b4ac189f91d2ebadfac99"} Dec 05 12:36:56 crc kubenswrapper[4807]: I1205 12:36:56.835635 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" podStartSLOduration=2.835604996 podStartE2EDuration="2.835604996s" podCreationTimestamp="2025-12-05 12:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:56.834163021 +0000 UTC m=+1846.328026290" watchObservedRunningTime="2025-12-05 12:36:56.835604996 +0000 UTC m=+1846.329468265" Dec 05 12:36:58 crc kubenswrapper[4807]: I1205 12:36:58.821143 4807 generic.go:334] "Generic (PLEG): container finished" podID="6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" containerID="dc3ff126ea0ecf2857481fe9694669ba2cd1d87a737b4ac189f91d2ebadfac99" exitCode=0 Dec 05 12:36:58 crc kubenswrapper[4807]: I1205 12:36:58.821422 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" event={"ID":"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2","Type":"ContainerDied","Data":"dc3ff126ea0ecf2857481fe9694669ba2cd1d87a737b4ac189f91d2ebadfac99"} Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.185030 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.378440 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-config-data\") pod \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.378631 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqxxj\" (UniqueName: \"kubernetes.io/projected/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-kube-api-access-wqxxj\") pod \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.378753 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-combined-ca-bundle\") pod \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.378777 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-db-sync-config-data\") pod \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\" (UID: \"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2\") " Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.385091 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" (UID: "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.385109 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-kube-api-access-wqxxj" (OuterVolumeSpecName: "kube-api-access-wqxxj") pod "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" (UID: "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2"). InnerVolumeSpecName "kube-api-access-wqxxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.401808 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" (UID: "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.420134 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-config-data" (OuterVolumeSpecName: "config-data") pod "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" (UID: "6cc0e2e2-2e0a-4092-9114-59bb2325f2c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.480172 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.480222 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqxxj\" (UniqueName: \"kubernetes.io/projected/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-kube-api-access-wqxxj\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.480237 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.480249 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.837018 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" event={"ID":"6cc0e2e2-2e0a-4092-9114-59bb2325f2c2","Type":"ContainerDied","Data":"f1e49a4da551e7e94fe7f55f31fb6f3a59346f3651a98accce879a61e812eb6b"} Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.837275 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1e49a4da551e7e94fe7f55f31fb6f3a59346f3651a98accce879a61e812eb6b" Dec 05 12:37:00 crc kubenswrapper[4807]: I1205 12:37:00.837063 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.448460 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:01 crc kubenswrapper[4807]: E1205 12:37:01.449777 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" containerName="watcher-kuttl-db-sync" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.449800 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" containerName="watcher-kuttl-db-sync" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.450131 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" containerName="watcher-kuttl-db-sync" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.451792 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.456811 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-z96pj" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.457348 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.469383 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.500966 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.505468 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.510288 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.512946 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7538542-0f1d-439b-ad97-dd795b74a2c8-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.512999 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513045 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513169 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513217 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z4fz\" (UniqueName: \"kubernetes.io/projected/a7538542-0f1d-439b-ad97-dd795b74a2c8-kube-api-access-8z4fz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513269 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513311 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513346 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec775590-73ae-4a01-89ec-63c94a9e6d58-logs\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513456 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xfrl\" (UniqueName: \"kubernetes.io/projected/ec775590-73ae-4a01-89ec-63c94a9e6d58-kube-api-access-7xfrl\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513490 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513587 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513677 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.513735 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.586629 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.588170 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.592445 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.598047 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627586 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xfrl\" (UniqueName: \"kubernetes.io/projected/ec775590-73ae-4a01-89ec-63c94a9e6d58-kube-api-access-7xfrl\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627657 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627714 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627743 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627804 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744b33e1-93da-433c-9abf-319ebc9c5e2e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627842 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627872 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzvbz\" (UniqueName: \"kubernetes.io/projected/744b33e1-93da-433c-9abf-319ebc9c5e2e-kube-api-access-jzvbz\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627905 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627931 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627967 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7538542-0f1d-439b-ad97-dd795b74a2c8-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.627993 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.628024 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.628055 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z4fz\" (UniqueName: \"kubernetes.io/projected/a7538542-0f1d-439b-ad97-dd795b74a2c8-kube-api-access-8z4fz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.628080 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.628107 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.628139 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.628166 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec775590-73ae-4a01-89ec-63c94a9e6d58-logs\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.628717 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec775590-73ae-4a01-89ec-63c94a9e6d58-logs\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.630014 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7538542-0f1d-439b-ad97-dd795b74a2c8-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.635202 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.635791 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.637515 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.639156 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.640700 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.641146 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.641628 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.649090 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.654203 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xfrl\" (UniqueName: \"kubernetes.io/projected/ec775590-73ae-4a01-89ec-63c94a9e6d58-kube-api-access-7xfrl\") pod \"watcher-kuttl-api-0\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.657173 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z4fz\" (UniqueName: \"kubernetes.io/projected/a7538542-0f1d-439b-ad97-dd795b74a2c8-kube-api-access-8z4fz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.729904 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.729966 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744b33e1-93da-433c-9abf-319ebc9c5e2e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.729990 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzvbz\" (UniqueName: \"kubernetes.io/projected/744b33e1-93da-433c-9abf-319ebc9c5e2e-kube-api-access-jzvbz\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.730013 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.730069 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.730426 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744b33e1-93da-433c-9abf-319ebc9c5e2e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.732929 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.733669 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.740103 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.753951 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzvbz\" (UniqueName: \"kubernetes.io/projected/744b33e1-93da-433c-9abf-319ebc9c5e2e-kube-api-access-jzvbz\") pod \"watcher-kuttl-applier-0\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.771879 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.831580 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:01 crc kubenswrapper[4807]: I1205 12:37:01.909560 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.236637 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:37:02 crc kubenswrapper[4807]: E1205 12:37:02.236937 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.287413 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:02 crc kubenswrapper[4807]: W1205 12:37:02.293590 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec775590_73ae_4a01_89ec_63c94a9e6d58.slice/crio-2d15dcba6193a54a9d102febdca0c0d61d461394713ca51351ae627253b32e16 WatchSource:0}: Error finding container 2d15dcba6193a54a9d102febdca0c0d61d461394713ca51351ae627253b32e16: Status 404 returned error can't find the container with id 2d15dcba6193a54a9d102febdca0c0d61d461394713ca51351ae627253b32e16 Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.397433 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:02 crc kubenswrapper[4807]: W1205 12:37:02.399136 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7538542_0f1d_439b_ad97_dd795b74a2c8.slice/crio-f9a8af81f6e669c788e770b0722e83fb0b33985ee4094e8a86404dd622938c91 WatchSource:0}: Error finding container f9a8af81f6e669c788e770b0722e83fb0b33985ee4094e8a86404dd622938c91: Status 404 returned error can't find the container with id f9a8af81f6e669c788e770b0722e83fb0b33985ee4094e8a86404dd622938c91 Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.475910 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lb95l"] Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.482288 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.514688 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb95l"] Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.522198 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.547496 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-utilities\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.547632 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tqjw\" (UniqueName: \"kubernetes.io/projected/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-kube-api-access-6tqjw\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.547693 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-catalog-content\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.649553 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-utilities\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.649786 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tqjw\" (UniqueName: \"kubernetes.io/projected/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-kube-api-access-6tqjw\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.649903 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-catalog-content\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.650541 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-catalog-content\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.650969 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-utilities\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.669883 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wc8n6"] Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.672013 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.674281 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tqjw\" (UniqueName: \"kubernetes.io/projected/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-kube-api-access-6tqjw\") pod \"redhat-marketplace-lb95l\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.690129 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wc8n6"] Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.752192 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-catalog-content\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.752272 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdg7f\" (UniqueName: \"kubernetes.io/projected/dded1516-37e3-4211-b616-48c53897c193-kube-api-access-gdg7f\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.752360 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-utilities\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.813816 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.856178 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-catalog-content\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.856301 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdg7f\" (UniqueName: \"kubernetes.io/projected/dded1516-37e3-4211-b616-48c53897c193-kube-api-access-gdg7f\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.856465 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-utilities\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.856720 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-catalog-content\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.856958 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-utilities\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.872714 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"744b33e1-93da-433c-9abf-319ebc9c5e2e","Type":"ContainerStarted","Data":"574bdf240545faefe5c75dbdf95b67ae92273949732a7e02fc2db72bc8b2f5e9"} Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.878729 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdg7f\" (UniqueName: \"kubernetes.io/projected/dded1516-37e3-4211-b616-48c53897c193-kube-api-access-gdg7f\") pod \"redhat-operators-wc8n6\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.881368 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"a7538542-0f1d-439b-ad97-dd795b74a2c8","Type":"ContainerStarted","Data":"1da0539b6a905359ed4fd2edb1da67e39f95eb7171ae351751fa4fcccdff58de"} Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.881409 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"a7538542-0f1d-439b-ad97-dd795b74a2c8","Type":"ContainerStarted","Data":"f9a8af81f6e669c788e770b0722e83fb0b33985ee4094e8a86404dd622938c91"} Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.884676 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"ec775590-73ae-4a01-89ec-63c94a9e6d58","Type":"ContainerStarted","Data":"9aa6453501e19036e109381141e6e0b6078c89c734e12740692434cc0fc58c7b"} Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.884956 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"ec775590-73ae-4a01-89ec-63c94a9e6d58","Type":"ContainerStarted","Data":"2d15dcba6193a54a9d102febdca0c0d61d461394713ca51351ae627253b32e16"} Dec 05 12:37:02 crc kubenswrapper[4807]: I1205 12:37:02.907379 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=1.90736219 podStartE2EDuration="1.90736219s" podCreationTimestamp="2025-12-05 12:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:02.906867388 +0000 UTC m=+1852.400730677" watchObservedRunningTime="2025-12-05 12:37:02.90736219 +0000 UTC m=+1852.401225459" Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.103827 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.391411 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb95l"] Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.675766 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wc8n6"] Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.924762 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8n6" event={"ID":"dded1516-37e3-4211-b616-48c53897c193","Type":"ContainerStarted","Data":"263faaf75be98f69c0ecdd3ad83971cdfd19720e650eb2a1aca3c5daa4ca2300"} Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.958822 4807 generic.go:334] "Generic (PLEG): container finished" podID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerID="86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465" exitCode=0 Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.958934 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb95l" event={"ID":"72ca4b5c-d10d-476b-9743-e1cd5b6b343a","Type":"ContainerDied","Data":"86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465"} Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.958966 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb95l" event={"ID":"72ca4b5c-d10d-476b-9743-e1cd5b6b343a","Type":"ContainerStarted","Data":"988a32e4402c95052562cf98f94bf7a1a8b0f9d0c65160f414012b78431ca898"} Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.989121 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"744b33e1-93da-433c-9abf-319ebc9c5e2e","Type":"ContainerStarted","Data":"cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff"} Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.994681 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"ec775590-73ae-4a01-89ec-63c94a9e6d58","Type":"ContainerStarted","Data":"8deae7b5488728b8122f886422109da9ba6568f6b7d39abcdf12ebf5df1dbb04"} Dec 05 12:37:03 crc kubenswrapper[4807]: I1205 12:37:03.994958 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:04 crc kubenswrapper[4807]: I1205 12:37:04.010958 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=3.01094069 podStartE2EDuration="3.01094069s" podCreationTimestamp="2025-12-05 12:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:04.009226428 +0000 UTC m=+1853.503089687" watchObservedRunningTime="2025-12-05 12:37:04.01094069 +0000 UTC m=+1853.504803969" Dec 05 12:37:04 crc kubenswrapper[4807]: I1205 12:37:04.030809 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=3.030792074 podStartE2EDuration="3.030792074s" podCreationTimestamp="2025-12-05 12:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:04.028760725 +0000 UTC m=+1853.522623994" watchObservedRunningTime="2025-12-05 12:37:04.030792074 +0000 UTC m=+1853.524655343" Dec 05 12:37:04 crc kubenswrapper[4807]: I1205 12:37:04.063629 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-db-create-22lp6"] Dec 05 12:37:04 crc kubenswrapper[4807]: I1205 12:37:04.085644 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc"] Dec 05 12:37:04 crc kubenswrapper[4807]: I1205 12:37:04.091736 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-fd9f-account-create-update-5cpvc"] Dec 05 12:37:04 crc kubenswrapper[4807]: I1205 12:37:04.097525 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-db-create-22lp6"] Dec 05 12:37:05 crc kubenswrapper[4807]: I1205 12:37:05.009463 4807 generic.go:334] "Generic (PLEG): container finished" podID="dded1516-37e3-4211-b616-48c53897c193" containerID="956ef428eb9306352a6ff491eaa80618cc9a1b2522465ec1658e8a0d62fdd510" exitCode=0 Dec 05 12:37:05 crc kubenswrapper[4807]: I1205 12:37:05.009546 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8n6" event={"ID":"dded1516-37e3-4211-b616-48c53897c193","Type":"ContainerDied","Data":"956ef428eb9306352a6ff491eaa80618cc9a1b2522465ec1658e8a0d62fdd510"} Dec 05 12:37:05 crc kubenswrapper[4807]: I1205 12:37:05.023173 4807 generic.go:334] "Generic (PLEG): container finished" podID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerID="2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16" exitCode=0 Dec 05 12:37:05 crc kubenswrapper[4807]: I1205 12:37:05.023233 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb95l" event={"ID":"72ca4b5c-d10d-476b-9743-e1cd5b6b343a","Type":"ContainerDied","Data":"2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16"} Dec 05 12:37:05 crc kubenswrapper[4807]: I1205 12:37:05.252737 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3d95952-eefb-4b02-8480-aa8624398190" path="/var/lib/kubelet/pods/a3d95952-eefb-4b02-8480-aa8624398190/volumes" Dec 05 12:37:05 crc kubenswrapper[4807]: I1205 12:37:05.254162 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1bef399-bb15-4310-a951-08b0ca8ef58a" path="/var/lib/kubelet/pods/d1bef399-bb15-4310-a951-08b0ca8ef58a/volumes" Dec 05 12:37:06 crc kubenswrapper[4807]: I1205 12:37:06.058713 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8n6" event={"ID":"dded1516-37e3-4211-b616-48c53897c193","Type":"ContainerStarted","Data":"e5e794f02874ba21c582162807c58ffb2d52f96f95f17b7c7c53db0e27ddb43a"} Dec 05 12:37:06 crc kubenswrapper[4807]: I1205 12:37:06.064457 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb95l" event={"ID":"72ca4b5c-d10d-476b-9743-e1cd5b6b343a","Type":"ContainerStarted","Data":"ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901"} Dec 05 12:37:06 crc kubenswrapper[4807]: I1205 12:37:06.108884 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lb95l" podStartSLOduration=2.650257333 podStartE2EDuration="4.108864778s" podCreationTimestamp="2025-12-05 12:37:02 +0000 UTC" firstStartedPulling="2025-12-05 12:37:03.966707581 +0000 UTC m=+1853.460570860" lastFinishedPulling="2025-12-05 12:37:05.425315026 +0000 UTC m=+1854.919178305" observedRunningTime="2025-12-05 12:37:06.100920655 +0000 UTC m=+1855.594783924" watchObservedRunningTime="2025-12-05 12:37:06.108864778 +0000 UTC m=+1855.602728047" Dec 05 12:37:06 crc kubenswrapper[4807]: I1205 12:37:06.742890 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:06 crc kubenswrapper[4807]: I1205 12:37:06.772567 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:06 crc kubenswrapper[4807]: I1205 12:37:06.909800 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:07 crc kubenswrapper[4807]: I1205 12:37:07.076773 4807 generic.go:334] "Generic (PLEG): container finished" podID="dded1516-37e3-4211-b616-48c53897c193" containerID="e5e794f02874ba21c582162807c58ffb2d52f96f95f17b7c7c53db0e27ddb43a" exitCode=0 Dec 05 12:37:07 crc kubenswrapper[4807]: I1205 12:37:07.077049 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8n6" event={"ID":"dded1516-37e3-4211-b616-48c53897c193","Type":"ContainerDied","Data":"e5e794f02874ba21c582162807c58ffb2d52f96f95f17b7c7c53db0e27ddb43a"} Dec 05 12:37:08 crc kubenswrapper[4807]: I1205 12:37:08.088133 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8n6" event={"ID":"dded1516-37e3-4211-b616-48c53897c193","Type":"ContainerStarted","Data":"76380f833e85f402ad67a9a682d95b6c710f0168cd6c75112b497eed72cd7081"} Dec 05 12:37:08 crc kubenswrapper[4807]: I1205 12:37:08.117916 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wc8n6" podStartSLOduration=3.629168666 podStartE2EDuration="6.11789519s" podCreationTimestamp="2025-12-05 12:37:02 +0000 UTC" firstStartedPulling="2025-12-05 12:37:05.019823933 +0000 UTC m=+1854.513687202" lastFinishedPulling="2025-12-05 12:37:07.508550457 +0000 UTC m=+1857.002413726" observedRunningTime="2025-12-05 12:37:08.1121762 +0000 UTC m=+1857.606039469" watchObservedRunningTime="2025-12-05 12:37:08.11789519 +0000 UTC m=+1857.611758459" Dec 05 12:37:11 crc kubenswrapper[4807]: I1205 12:37:11.773377 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:11 crc kubenswrapper[4807]: I1205 12:37:11.779340 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:11 crc kubenswrapper[4807]: I1205 12:37:11.832087 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:11 crc kubenswrapper[4807]: I1205 12:37:11.869215 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:11 crc kubenswrapper[4807]: I1205 12:37:11.909888 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:11 crc kubenswrapper[4807]: I1205 12:37:11.934439 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:12 crc kubenswrapper[4807]: I1205 12:37:12.117720 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:12 crc kubenswrapper[4807]: I1205 12:37:12.124256 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:12 crc kubenswrapper[4807]: I1205 12:37:12.143162 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:12 crc kubenswrapper[4807]: I1205 12:37:12.152402 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:12 crc kubenswrapper[4807]: I1205 12:37:12.814513 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:12 crc kubenswrapper[4807]: I1205 12:37:12.814574 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:12 crc kubenswrapper[4807]: I1205 12:37:12.872459 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:13 crc kubenswrapper[4807]: I1205 12:37:13.104666 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:13 crc kubenswrapper[4807]: I1205 12:37:13.104726 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:13 crc kubenswrapper[4807]: I1205 12:37:13.147597 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:13 crc kubenswrapper[4807]: I1205 12:37:13.179699 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:13 crc kubenswrapper[4807]: I1205 12:37:13.200057 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:16 crc kubenswrapper[4807]: I1205 12:37:16.509374 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:16 crc kubenswrapper[4807]: I1205 12:37:16.509947 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-central-agent" containerID="cri-o://b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af" gracePeriod=30 Dec 05 12:37:16 crc kubenswrapper[4807]: I1205 12:37:16.510103 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="proxy-httpd" containerID="cri-o://5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342" gracePeriod=30 Dec 05 12:37:16 crc kubenswrapper[4807]: I1205 12:37:16.510164 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="sg-core" containerID="cri-o://15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9" gracePeriod=30 Dec 05 12:37:16 crc kubenswrapper[4807]: I1205 12:37:16.510198 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-notification-agent" containerID="cri-o://d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b" gracePeriod=30 Dec 05 12:37:16 crc kubenswrapper[4807]: I1205 12:37:16.525682 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.169496 4807 generic.go:334] "Generic (PLEG): container finished" podID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerID="5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342" exitCode=0 Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.169827 4807 generic.go:334] "Generic (PLEG): container finished" podID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerID="15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9" exitCode=2 Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.169839 4807 generic.go:334] "Generic (PLEG): container finished" podID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerID="b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af" exitCode=0 Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.169553 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerDied","Data":"5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342"} Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.169872 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerDied","Data":"15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9"} Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.169885 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerDied","Data":"b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af"} Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.236162 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:37:17 crc kubenswrapper[4807]: E1205 12:37:17.236621 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.462006 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb95l"] Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.462263 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lb95l" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="registry-server" containerID="cri-o://ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901" gracePeriod=2 Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.873869 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wc8n6"] Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.874470 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wc8n6" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="registry-server" containerID="cri-o://76380f833e85f402ad67a9a682d95b6c710f0168cd6c75112b497eed72cd7081" gracePeriod=2 Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.918092 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.954204 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-utilities\") pod \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.954399 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tqjw\" (UniqueName: \"kubernetes.io/projected/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-kube-api-access-6tqjw\") pod \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.954446 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-catalog-content\") pod \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\" (UID: \"72ca4b5c-d10d-476b-9743-e1cd5b6b343a\") " Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.956148 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-utilities" (OuterVolumeSpecName: "utilities") pod "72ca4b5c-d10d-476b-9743-e1cd5b6b343a" (UID: "72ca4b5c-d10d-476b-9743-e1cd5b6b343a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.965712 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-kube-api-access-6tqjw" (OuterVolumeSpecName: "kube-api-access-6tqjw") pod "72ca4b5c-d10d-476b-9743-e1cd5b6b343a" (UID: "72ca4b5c-d10d-476b-9743-e1cd5b6b343a"). InnerVolumeSpecName "kube-api-access-6tqjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:17 crc kubenswrapper[4807]: I1205 12:37:17.977986 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72ca4b5c-d10d-476b-9743-e1cd5b6b343a" (UID: "72ca4b5c-d10d-476b-9743-e1cd5b6b343a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.056954 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tqjw\" (UniqueName: \"kubernetes.io/projected/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-kube-api-access-6tqjw\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.057008 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.057023 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ca4b5c-d10d-476b-9743-e1cd5b6b343a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.111969 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.205:3000/\": dial tcp 10.217.0.205:3000: connect: connection refused" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.183188 4807 generic.go:334] "Generic (PLEG): container finished" podID="dded1516-37e3-4211-b616-48c53897c193" containerID="76380f833e85f402ad67a9a682d95b6c710f0168cd6c75112b497eed72cd7081" exitCode=0 Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.183269 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8n6" event={"ID":"dded1516-37e3-4211-b616-48c53897c193","Type":"ContainerDied","Data":"76380f833e85f402ad67a9a682d95b6c710f0168cd6c75112b497eed72cd7081"} Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.185429 4807 generic.go:334] "Generic (PLEG): container finished" podID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerID="ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901" exitCode=0 Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.185461 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb95l" event={"ID":"72ca4b5c-d10d-476b-9743-e1cd5b6b343a","Type":"ContainerDied","Data":"ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901"} Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.185482 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb95l" event={"ID":"72ca4b5c-d10d-476b-9743-e1cd5b6b343a","Type":"ContainerDied","Data":"988a32e4402c95052562cf98f94bf7a1a8b0f9d0c65160f414012b78431ca898"} Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.185504 4807 scope.go:117] "RemoveContainer" containerID="ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.185671 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb95l" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.224855 4807 scope.go:117] "RemoveContainer" containerID="2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.230883 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb95l"] Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.241397 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb95l"] Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.256737 4807 scope.go:117] "RemoveContainer" containerID="86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.293663 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.298105 4807 scope.go:117] "RemoveContainer" containerID="ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901" Dec 05 12:37:18 crc kubenswrapper[4807]: E1205 12:37:18.298399 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901\": container with ID starting with ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901 not found: ID does not exist" containerID="ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.298700 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901"} err="failed to get container status \"ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901\": rpc error: code = NotFound desc = could not find container \"ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901\": container with ID starting with ce5a4ec060111c4e96bc28cd94f92dd2b6de021c88ea5447e79c7889e2bf2901 not found: ID does not exist" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.298727 4807 scope.go:117] "RemoveContainer" containerID="2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16" Dec 05 12:37:18 crc kubenswrapper[4807]: E1205 12:37:18.299379 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16\": container with ID starting with 2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16 not found: ID does not exist" containerID="2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.299430 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16"} err="failed to get container status \"2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16\": rpc error: code = NotFound desc = could not find container \"2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16\": container with ID starting with 2b48107f3bae1c6b1f7db38346a3be878484837de2df29191ba9a20bad980b16 not found: ID does not exist" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.299461 4807 scope.go:117] "RemoveContainer" containerID="86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465" Dec 05 12:37:18 crc kubenswrapper[4807]: E1205 12:37:18.299798 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465\": container with ID starting with 86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465 not found: ID does not exist" containerID="86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.299824 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465"} err="failed to get container status \"86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465\": rpc error: code = NotFound desc = could not find container \"86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465\": container with ID starting with 86cb5769655ab35b4a9d453cd2bea8b96ac88224e30cdcb018a566a105192465 not found: ID does not exist" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.361803 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdg7f\" (UniqueName: \"kubernetes.io/projected/dded1516-37e3-4211-b616-48c53897c193-kube-api-access-gdg7f\") pod \"dded1516-37e3-4211-b616-48c53897c193\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.362110 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-catalog-content\") pod \"dded1516-37e3-4211-b616-48c53897c193\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.362322 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-utilities\") pod \"dded1516-37e3-4211-b616-48c53897c193\" (UID: \"dded1516-37e3-4211-b616-48c53897c193\") " Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.366628 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-utilities" (OuterVolumeSpecName: "utilities") pod "dded1516-37e3-4211-b616-48c53897c193" (UID: "dded1516-37e3-4211-b616-48c53897c193"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.368286 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dded1516-37e3-4211-b616-48c53897c193-kube-api-access-gdg7f" (OuterVolumeSpecName: "kube-api-access-gdg7f") pod "dded1516-37e3-4211-b616-48c53897c193" (UID: "dded1516-37e3-4211-b616-48c53897c193"). InnerVolumeSpecName "kube-api-access-gdg7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.464131 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdg7f\" (UniqueName: \"kubernetes.io/projected/dded1516-37e3-4211-b616-48c53897c193-kube-api-access-gdg7f\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.464181 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.483997 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dded1516-37e3-4211-b616-48c53897c193" (UID: "dded1516-37e3-4211-b616-48c53897c193"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:18 crc kubenswrapper[4807]: I1205 12:37:18.565109 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dded1516-37e3-4211-b616-48c53897c193-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.194783 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8n6" event={"ID":"dded1516-37e3-4211-b616-48c53897c193","Type":"ContainerDied","Data":"263faaf75be98f69c0ecdd3ad83971cdfd19720e650eb2a1aca3c5daa4ca2300"} Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.195184 4807 scope.go:117] "RemoveContainer" containerID="76380f833e85f402ad67a9a682d95b6c710f0168cd6c75112b497eed72cd7081" Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.194819 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8n6" Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.216451 4807 scope.go:117] "RemoveContainer" containerID="e5e794f02874ba21c582162807c58ffb2d52f96f95f17b7c7c53db0e27ddb43a" Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.242747 4807 scope.go:117] "RemoveContainer" containerID="956ef428eb9306352a6ff491eaa80618cc9a1b2522465ec1658e8a0d62fdd510" Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.250693 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" path="/var/lib/kubelet/pods/72ca4b5c-d10d-476b-9743-e1cd5b6b343a/volumes" Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.344619 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wc8n6"] Dec 05 12:37:19 crc kubenswrapper[4807]: I1205 12:37:19.351229 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wc8n6"] Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.248469 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dded1516-37e3-4211-b616-48c53897c193" path="/var/lib/kubelet/pods/dded1516-37e3-4211-b616-48c53897c193/volumes" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.697501 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm"] Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.707159 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vj6rm"] Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733097 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcherfc03-account-delete-6ljgk"] Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.733502 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="extract-content" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733540 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="extract-content" Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.733569 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="extract-utilities" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733576 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="extract-utilities" Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.733588 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="registry-server" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733593 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="registry-server" Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.733602 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="registry-server" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733608 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="registry-server" Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.733623 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="extract-utilities" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733629 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="extract-utilities" Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.733647 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="extract-content" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733653 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="extract-content" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733821 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dded1516-37e3-4211-b616-48c53897c193" containerName="registry-server" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.733842 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ca4b5c-d10d-476b-9743-e1cd5b6b343a" containerName="registry-server" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.734662 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.743036 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.743486 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="744b33e1-93da-433c-9abf-319ebc9c5e2e" containerName="watcher-applier" containerID="cri-o://cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" gracePeriod=30 Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.751406 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherfc03-account-delete-6ljgk"] Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.826661 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.826891 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-kuttl-api-log" containerID="cri-o://9aa6453501e19036e109381141e6e0b6078c89c734e12740692434cc0fc58c7b" gracePeriod=30 Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.827005 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-api" containerID="cri-o://8deae7b5488728b8122f886422109da9ba6568f6b7d39abcdf12ebf5df1dbb04" gracePeriod=30 Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.828326 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq4tk\" (UniqueName: \"kubernetes.io/projected/51f334cd-81eb-447e-a00f-147c84da6027-kube-api-access-zq4tk\") pod \"watcherfc03-account-delete-6ljgk\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.828495 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f334cd-81eb-447e-a00f-147c84da6027-operator-scripts\") pod \"watcherfc03-account-delete-6ljgk\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.873201 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.873447 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="a7538542-0f1d-439b-ad97-dd795b74a2c8" containerName="watcher-decision-engine" containerID="cri-o://1da0539b6a905359ed4fd2edb1da67e39f95eb7171ae351751fa4fcccdff58de" gracePeriod=30 Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.909977 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.913727 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.915370 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.916827 4807 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 05 12:37:21 crc kubenswrapper[4807]: E1205 12:37:21.916867 4807 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="744b33e1-93da-433c-9abf-319ebc9c5e2e" containerName="watcher-applier" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.929960 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f334cd-81eb-447e-a00f-147c84da6027-operator-scripts\") pod \"watcherfc03-account-delete-6ljgk\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.930052 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq4tk\" (UniqueName: \"kubernetes.io/projected/51f334cd-81eb-447e-a00f-147c84da6027-kube-api-access-zq4tk\") pod \"watcherfc03-account-delete-6ljgk\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.931063 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f334cd-81eb-447e-a00f-147c84da6027-operator-scripts\") pod \"watcherfc03-account-delete-6ljgk\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:21 crc kubenswrapper[4807]: I1205 12:37:21.962309 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq4tk\" (UniqueName: \"kubernetes.io/projected/51f334cd-81eb-447e-a00f-147c84da6027-kube-api-access-zq4tk\") pod \"watcherfc03-account-delete-6ljgk\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.031906 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-sg-core-conf-yaml\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.032029 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-ceilometer-tls-certs\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.032105 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-config-data\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.032153 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2lc8\" (UniqueName: \"kubernetes.io/projected/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-kube-api-access-l2lc8\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.032202 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-scripts\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.032228 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-log-httpd\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.032266 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-combined-ca-bundle\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.032293 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-run-httpd\") pod \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\" (UID: \"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4\") " Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.033063 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.033106 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.035453 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-scripts" (OuterVolumeSpecName: "scripts") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.037956 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-kube-api-access-l2lc8" (OuterVolumeSpecName: "kube-api-access-l2lc8") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "kube-api-access-l2lc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.061395 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.066617 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.088712 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.136387 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.136423 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2lc8\" (UniqueName: \"kubernetes.io/projected/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-kube-api-access-l2lc8\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.136435 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.136444 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.136452 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.136620 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.140052 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.149024 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-config-data" (OuterVolumeSpecName: "config-data") pod "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" (UID: "ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.238173 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.238424 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.255229 4807 generic.go:334] "Generic (PLEG): container finished" podID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerID="d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b" exitCode=0 Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.255289 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerDied","Data":"d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b"} Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.255315 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4","Type":"ContainerDied","Data":"acf0a0864d53847eb5014f2fec8b2300ead0baad6605d033d6115785dc94b40a"} Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.255331 4807 scope.go:117] "RemoveContainer" containerID="5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.255460 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.260397 4807 generic.go:334] "Generic (PLEG): container finished" podID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerID="9aa6453501e19036e109381141e6e0b6078c89c734e12740692434cc0fc58c7b" exitCode=143 Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.260479 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"ec775590-73ae-4a01-89ec-63c94a9e6d58","Type":"ContainerDied","Data":"9aa6453501e19036e109381141e6e0b6078c89c734e12740692434cc0fc58c7b"} Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.281713 4807 scope.go:117] "RemoveContainer" containerID="15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.311257 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.329972 4807 scope.go:117] "RemoveContainer" containerID="d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.342245 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.357778 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.358197 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-notification-agent" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358225 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-notification-agent" Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.358242 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="sg-core" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358251 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="sg-core" Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.358267 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="proxy-httpd" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358276 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="proxy-httpd" Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.358301 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-central-agent" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358309 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-central-agent" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358875 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="sg-core" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358905 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-central-agent" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358917 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="proxy-httpd" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.358930 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" containerName="ceilometer-notification-agent" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.361635 4807 scope.go:117] "RemoveContainer" containerID="b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.363544 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.366778 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.366938 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.370102 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.370350 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.387935 4807 scope.go:117] "RemoveContainer" containerID="5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342" Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.396596 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342\": container with ID starting with 5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342 not found: ID does not exist" containerID="5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.396639 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342"} err="failed to get container status \"5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342\": rpc error: code = NotFound desc = could not find container \"5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342\": container with ID starting with 5fc6251b8eb8129acfe27ded270b96f34b3080b509b85fb08389be06a8589342 not found: ID does not exist" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.396662 4807 scope.go:117] "RemoveContainer" containerID="15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9" Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.399425 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9\": container with ID starting with 15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9 not found: ID does not exist" containerID="15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.399467 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9"} err="failed to get container status \"15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9\": rpc error: code = NotFound desc = could not find container \"15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9\": container with ID starting with 15ecceb2650bb08bee78d9ed77b294e942eca6f04fb7bfdaa99ded318bcb40f9 not found: ID does not exist" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.399511 4807 scope.go:117] "RemoveContainer" containerID="d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.400131 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherfc03-account-delete-6ljgk"] Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.400500 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b\": container with ID starting with d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b not found: ID does not exist" containerID="d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.400696 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b"} err="failed to get container status \"d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b\": rpc error: code = NotFound desc = could not find container \"d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b\": container with ID starting with d1f0a75d12954d195cce23fb974493b0171a2f894b38a469f2cfe25e1e8b866b not found: ID does not exist" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.400783 4807 scope.go:117] "RemoveContainer" containerID="b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af" Dec 05 12:37:22 crc kubenswrapper[4807]: E1205 12:37:22.401357 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af\": container with ID starting with b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af not found: ID does not exist" containerID="b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.401610 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af"} err="failed to get container status \"b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af\": rpc error: code = NotFound desc = could not find container \"b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af\": container with ID starting with b3a5b76eba0afed041a00d6882049a25f6fc6920b146b7e58460e7de7a5834af not found: ID does not exist" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442055 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442114 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ngxg\" (UniqueName: \"kubernetes.io/projected/2f7edd43-2642-40a1-8fc9-70ee2daf650f-kube-api-access-9ngxg\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442160 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-log-httpd\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442223 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442263 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-config-data\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442323 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-scripts\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442352 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.442374 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-run-httpd\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.543794 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ngxg\" (UniqueName: \"kubernetes.io/projected/2f7edd43-2642-40a1-8fc9-70ee2daf650f-kube-api-access-9ngxg\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544190 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-log-httpd\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544324 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544429 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-config-data\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544586 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-scripts\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544678 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544779 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-run-httpd\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544894 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.546159 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-run-httpd\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.544695 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-log-httpd\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.550692 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-scripts\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.551968 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-config-data\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.553696 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.554466 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.561145 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.568173 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ngxg\" (UniqueName: \"kubernetes.io/projected/2f7edd43-2642-40a1-8fc9-70ee2daf650f-kube-api-access-9ngxg\") pod \"ceilometer-0\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:22 crc kubenswrapper[4807]: I1205 12:37:22.679423 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.193657 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:23 crc kubenswrapper[4807]: W1205 12:37:23.196917 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f7edd43_2642_40a1_8fc9_70ee2daf650f.slice/crio-b4abaead45f910316db7dadd6bc9ea2b9b9e8d80cb3d9d5023d7558721c296a7 WatchSource:0}: Error finding container b4abaead45f910316db7dadd6bc9ea2b9b9e8d80cb3d9d5023d7558721c296a7: Status 404 returned error can't find the container with id b4abaead45f910316db7dadd6bc9ea2b9b9e8d80cb3d9d5023d7558721c296a7 Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.254396 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cc0e2e2-2e0a-4092-9114-59bb2325f2c2" path="/var/lib/kubelet/pods/6cc0e2e2-2e0a-4092-9114-59bb2325f2c2/volumes" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.256810 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4" path="/var/lib/kubelet/pods/ebda6f78-25bc-4bb4-bedd-d5c7a7285bd4/volumes" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.287675 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerStarted","Data":"b4abaead45f910316db7dadd6bc9ea2b9b9e8d80cb3d9d5023d7558721c296a7"} Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.292897 4807 generic.go:334] "Generic (PLEG): container finished" podID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerID="8deae7b5488728b8122f886422109da9ba6568f6b7d39abcdf12ebf5df1dbb04" exitCode=0 Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.292966 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"ec775590-73ae-4a01-89ec-63c94a9e6d58","Type":"ContainerDied","Data":"8deae7b5488728b8122f886422109da9ba6568f6b7d39abcdf12ebf5df1dbb04"} Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.294461 4807 generic.go:334] "Generic (PLEG): container finished" podID="51f334cd-81eb-447e-a00f-147c84da6027" containerID="d9f118d16acd2df0531590783830463dbfd4d4d5e6003d5fd7df3399ba195ec7" exitCode=0 Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.294509 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" event={"ID":"51f334cd-81eb-447e-a00f-147c84da6027","Type":"ContainerDied","Data":"d9f118d16acd2df0531590783830463dbfd4d4d5e6003d5fd7df3399ba195ec7"} Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.294748 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" event={"ID":"51f334cd-81eb-447e-a00f-147c84da6027","Type":"ContainerStarted","Data":"ef6f2f79f34a37274a10c136d0357e03b99dbc5419007d13e4eb7cad2effe32d"} Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.384553 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.567293 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec775590-73ae-4a01-89ec-63c94a9e6d58-logs\") pod \"ec775590-73ae-4a01-89ec-63c94a9e6d58\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.567335 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-custom-prometheus-ca\") pod \"ec775590-73ae-4a01-89ec-63c94a9e6d58\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.567400 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xfrl\" (UniqueName: \"kubernetes.io/projected/ec775590-73ae-4a01-89ec-63c94a9e6d58-kube-api-access-7xfrl\") pod \"ec775590-73ae-4a01-89ec-63c94a9e6d58\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.567470 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-config-data\") pod \"ec775590-73ae-4a01-89ec-63c94a9e6d58\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.567595 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-combined-ca-bundle\") pod \"ec775590-73ae-4a01-89ec-63c94a9e6d58\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.567650 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-cert-memcached-mtls\") pod \"ec775590-73ae-4a01-89ec-63c94a9e6d58\" (UID: \"ec775590-73ae-4a01-89ec-63c94a9e6d58\") " Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.567886 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec775590-73ae-4a01-89ec-63c94a9e6d58-logs" (OuterVolumeSpecName: "logs") pod "ec775590-73ae-4a01-89ec-63c94a9e6d58" (UID: "ec775590-73ae-4a01-89ec-63c94a9e6d58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.568518 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec775590-73ae-4a01-89ec-63c94a9e6d58-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.575937 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec775590-73ae-4a01-89ec-63c94a9e6d58-kube-api-access-7xfrl" (OuterVolumeSpecName: "kube-api-access-7xfrl") pod "ec775590-73ae-4a01-89ec-63c94a9e6d58" (UID: "ec775590-73ae-4a01-89ec-63c94a9e6d58"). InnerVolumeSpecName "kube-api-access-7xfrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.599562 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec775590-73ae-4a01-89ec-63c94a9e6d58" (UID: "ec775590-73ae-4a01-89ec-63c94a9e6d58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.607793 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "ec775590-73ae-4a01-89ec-63c94a9e6d58" (UID: "ec775590-73ae-4a01-89ec-63c94a9e6d58"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.617327 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-config-data" (OuterVolumeSpecName: "config-data") pod "ec775590-73ae-4a01-89ec-63c94a9e6d58" (UID: "ec775590-73ae-4a01-89ec-63c94a9e6d58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.651536 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "ec775590-73ae-4a01-89ec-63c94a9e6d58" (UID: "ec775590-73ae-4a01-89ec-63c94a9e6d58"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.686349 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.686391 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.686400 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.686430 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xfrl\" (UniqueName: \"kubernetes.io/projected/ec775590-73ae-4a01-89ec-63c94a9e6d58-kube-api-access-7xfrl\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:23 crc kubenswrapper[4807]: I1205 12:37:23.686442 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec775590-73ae-4a01-89ec-63c94a9e6d58-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.313471 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"ec775590-73ae-4a01-89ec-63c94a9e6d58","Type":"ContainerDied","Data":"2d15dcba6193a54a9d102febdca0c0d61d461394713ca51351ae627253b32e16"} Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.313486 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.313806 4807 scope.go:117] "RemoveContainer" containerID="8deae7b5488728b8122f886422109da9ba6568f6b7d39abcdf12ebf5df1dbb04" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.315054 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerStarted","Data":"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d"} Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.341646 4807 scope.go:117] "RemoveContainer" containerID="9aa6453501e19036e109381141e6e0b6078c89c734e12740692434cc0fc58c7b" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.366683 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.379374 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.391695 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.746773 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.804772 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq4tk\" (UniqueName: \"kubernetes.io/projected/51f334cd-81eb-447e-a00f-147c84da6027-kube-api-access-zq4tk\") pod \"51f334cd-81eb-447e-a00f-147c84da6027\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.804826 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f334cd-81eb-447e-a00f-147c84da6027-operator-scripts\") pod \"51f334cd-81eb-447e-a00f-147c84da6027\" (UID: \"51f334cd-81eb-447e-a00f-147c84da6027\") " Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.805277 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51f334cd-81eb-447e-a00f-147c84da6027-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51f334cd-81eb-447e-a00f-147c84da6027" (UID: "51f334cd-81eb-447e-a00f-147c84da6027"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.809551 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51f334cd-81eb-447e-a00f-147c84da6027-kube-api-access-zq4tk" (OuterVolumeSpecName: "kube-api-access-zq4tk") pod "51f334cd-81eb-447e-a00f-147c84da6027" (UID: "51f334cd-81eb-447e-a00f-147c84da6027"). InnerVolumeSpecName "kube-api-access-zq4tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.883002 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906034 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-combined-ca-bundle\") pod \"744b33e1-93da-433c-9abf-319ebc9c5e2e\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906116 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744b33e1-93da-433c-9abf-319ebc9c5e2e-logs\") pod \"744b33e1-93da-433c-9abf-319ebc9c5e2e\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906307 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzvbz\" (UniqueName: \"kubernetes.io/projected/744b33e1-93da-433c-9abf-319ebc9c5e2e-kube-api-access-jzvbz\") pod \"744b33e1-93da-433c-9abf-319ebc9c5e2e\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906337 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-cert-memcached-mtls\") pod \"744b33e1-93da-433c-9abf-319ebc9c5e2e\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906376 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-config-data\") pod \"744b33e1-93da-433c-9abf-319ebc9c5e2e\" (UID: \"744b33e1-93da-433c-9abf-319ebc9c5e2e\") " Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906672 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/744b33e1-93da-433c-9abf-319ebc9c5e2e-logs" (OuterVolumeSpecName: "logs") pod "744b33e1-93da-433c-9abf-319ebc9c5e2e" (UID: "744b33e1-93da-433c-9abf-319ebc9c5e2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906938 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744b33e1-93da-433c-9abf-319ebc9c5e2e-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906956 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq4tk\" (UniqueName: \"kubernetes.io/projected/51f334cd-81eb-447e-a00f-147c84da6027-kube-api-access-zq4tk\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.906978 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f334cd-81eb-447e-a00f-147c84da6027-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.910792 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/744b33e1-93da-433c-9abf-319ebc9c5e2e-kube-api-access-jzvbz" (OuterVolumeSpecName: "kube-api-access-jzvbz") pod "744b33e1-93da-433c-9abf-319ebc9c5e2e" (UID: "744b33e1-93da-433c-9abf-319ebc9c5e2e"). InnerVolumeSpecName "kube-api-access-jzvbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.943881 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "744b33e1-93da-433c-9abf-319ebc9c5e2e" (UID: "744b33e1-93da-433c-9abf-319ebc9c5e2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.971400 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-config-data" (OuterVolumeSpecName: "config-data") pod "744b33e1-93da-433c-9abf-319ebc9c5e2e" (UID: "744b33e1-93da-433c-9abf-319ebc9c5e2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4807]: I1205 12:37:24.973808 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "744b33e1-93da-433c-9abf-319ebc9c5e2e" (UID: "744b33e1-93da-433c-9abf-319ebc9c5e2e"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.008677 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzvbz\" (UniqueName: \"kubernetes.io/projected/744b33e1-93da-433c-9abf-319ebc9c5e2e-kube-api-access-jzvbz\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.008712 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.008721 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.008732 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/744b33e1-93da-433c-9abf-319ebc9c5e2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.248428 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" path="/var/lib/kubelet/pods/ec775590-73ae-4a01-89ec-63c94a9e6d58/volumes" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.336767 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerStarted","Data":"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48"} Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.340799 4807 generic.go:334] "Generic (PLEG): container finished" podID="744b33e1-93da-433c-9abf-319ebc9c5e2e" containerID="cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" exitCode=0 Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.340857 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"744b33e1-93da-433c-9abf-319ebc9c5e2e","Type":"ContainerDied","Data":"cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff"} Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.340879 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"744b33e1-93da-433c-9abf-319ebc9c5e2e","Type":"ContainerDied","Data":"574bdf240545faefe5c75dbdf95b67ae92273949732a7e02fc2db72bc8b2f5e9"} Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.340945 4807 scope.go:117] "RemoveContainer" containerID="cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.341069 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.346794 4807 generic.go:334] "Generic (PLEG): container finished" podID="a7538542-0f1d-439b-ad97-dd795b74a2c8" containerID="1da0539b6a905359ed4fd2edb1da67e39f95eb7171ae351751fa4fcccdff58de" exitCode=0 Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.346891 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"a7538542-0f1d-439b-ad97-dd795b74a2c8","Type":"ContainerDied","Data":"1da0539b6a905359ed4fd2edb1da67e39f95eb7171ae351751fa4fcccdff58de"} Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.350390 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" event={"ID":"51f334cd-81eb-447e-a00f-147c84da6027","Type":"ContainerDied","Data":"ef6f2f79f34a37274a10c136d0357e03b99dbc5419007d13e4eb7cad2effe32d"} Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.350437 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef6f2f79f34a37274a10c136d0357e03b99dbc5419007d13e4eb7cad2effe32d" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.350501 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherfc03-account-delete-6ljgk" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.372634 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.382575 4807 scope.go:117] "RemoveContainer" containerID="cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" Dec 05 12:37:25 crc kubenswrapper[4807]: E1205 12:37:25.383053 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff\": container with ID starting with cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff not found: ID does not exist" containerID="cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.383093 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff"} err="failed to get container status \"cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff\": rpc error: code = NotFound desc = could not find container \"cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff\": container with ID starting with cce8e54b63a732a33fd1840796a0b77dd72980ef55b715c242914420972143ff not found: ID does not exist" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.383813 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.422313 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.518843 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-combined-ca-bundle\") pod \"a7538542-0f1d-439b-ad97-dd795b74a2c8\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.519233 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-cert-memcached-mtls\") pod \"a7538542-0f1d-439b-ad97-dd795b74a2c8\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.519279 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-config-data\") pod \"a7538542-0f1d-439b-ad97-dd795b74a2c8\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.519303 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-custom-prometheus-ca\") pod \"a7538542-0f1d-439b-ad97-dd795b74a2c8\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.519336 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7538542-0f1d-439b-ad97-dd795b74a2c8-logs\") pod \"a7538542-0f1d-439b-ad97-dd795b74a2c8\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.520085 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z4fz\" (UniqueName: \"kubernetes.io/projected/a7538542-0f1d-439b-ad97-dd795b74a2c8-kube-api-access-8z4fz\") pod \"a7538542-0f1d-439b-ad97-dd795b74a2c8\" (UID: \"a7538542-0f1d-439b-ad97-dd795b74a2c8\") " Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.521920 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7538542-0f1d-439b-ad97-dd795b74a2c8-logs" (OuterVolumeSpecName: "logs") pod "a7538542-0f1d-439b-ad97-dd795b74a2c8" (UID: "a7538542-0f1d-439b-ad97-dd795b74a2c8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.536694 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7538542-0f1d-439b-ad97-dd795b74a2c8-kube-api-access-8z4fz" (OuterVolumeSpecName: "kube-api-access-8z4fz") pod "a7538542-0f1d-439b-ad97-dd795b74a2c8" (UID: "a7538542-0f1d-439b-ad97-dd795b74a2c8"). InnerVolumeSpecName "kube-api-access-8z4fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.557750 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7538542-0f1d-439b-ad97-dd795b74a2c8" (UID: "a7538542-0f1d-439b-ad97-dd795b74a2c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.580503 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "a7538542-0f1d-439b-ad97-dd795b74a2c8" (UID: "a7538542-0f1d-439b-ad97-dd795b74a2c8"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.588881 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-config-data" (OuterVolumeSpecName: "config-data") pod "a7538542-0f1d-439b-ad97-dd795b74a2c8" (UID: "a7538542-0f1d-439b-ad97-dd795b74a2c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.608367 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "a7538542-0f1d-439b-ad97-dd795b74a2c8" (UID: "a7538542-0f1d-439b-ad97-dd795b74a2c8"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.622344 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.622383 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.622391 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.622401 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7538542-0f1d-439b-ad97-dd795b74a2c8-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.622411 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z4fz\" (UniqueName: \"kubernetes.io/projected/a7538542-0f1d-439b-ad97-dd795b74a2c8-kube-api-access-8z4fz\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4807]: I1205 12:37:25.622422 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7538542-0f1d-439b-ad97-dd795b74a2c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.361754 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"a7538542-0f1d-439b-ad97-dd795b74a2c8","Type":"ContainerDied","Data":"f9a8af81f6e669c788e770b0722e83fb0b33985ee4094e8a86404dd622938c91"} Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.361813 4807 scope.go:117] "RemoveContainer" containerID="1da0539b6a905359ed4fd2edb1da67e39f95eb7171ae351751fa4fcccdff58de" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.361945 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.376159 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerStarted","Data":"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b"} Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.419601 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.428034 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.736446 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vjbb7"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.744433 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vjbb7"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.782450 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcherfc03-account-delete-6ljgk"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.794611 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.795315 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcherfc03-account-delete-6ljgk"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.802032 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-fc03-account-create-update-7fl7q"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836138 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-7gt2n"] Dec 05 12:37:26 crc kubenswrapper[4807]: E1205 12:37:26.836513 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-api" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836545 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-api" Dec 05 12:37:26 crc kubenswrapper[4807]: E1205 12:37:26.836565 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51f334cd-81eb-447e-a00f-147c84da6027" containerName="mariadb-account-delete" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836571 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="51f334cd-81eb-447e-a00f-147c84da6027" containerName="mariadb-account-delete" Dec 05 12:37:26 crc kubenswrapper[4807]: E1205 12:37:26.836591 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="744b33e1-93da-433c-9abf-319ebc9c5e2e" containerName="watcher-applier" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836598 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="744b33e1-93da-433c-9abf-319ebc9c5e2e" containerName="watcher-applier" Dec 05 12:37:26 crc kubenswrapper[4807]: E1205 12:37:26.836613 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7538542-0f1d-439b-ad97-dd795b74a2c8" containerName="watcher-decision-engine" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836619 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7538542-0f1d-439b-ad97-dd795b74a2c8" containerName="watcher-decision-engine" Dec 05 12:37:26 crc kubenswrapper[4807]: E1205 12:37:26.836628 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-kuttl-api-log" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836634 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-kuttl-api-log" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836798 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="744b33e1-93da-433c-9abf-319ebc9c5e2e" containerName="watcher-applier" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836820 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7538542-0f1d-439b-ad97-dd795b74a2c8" containerName="watcher-decision-engine" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836834 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-kuttl-api-log" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836844 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec775590-73ae-4a01-89ec-63c94a9e6d58" containerName="watcher-api" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.836853 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="51f334cd-81eb-447e-a00f-147c84da6027" containerName="mariadb-account-delete" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.837377 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.841325 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6jwg\" (UniqueName: \"kubernetes.io/projected/68413519-afb4-4f45-97e0-54b989b13fd6-kube-api-access-v6jwg\") pod \"watcher-db-create-7gt2n\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.841404 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68413519-afb4-4f45-97e0-54b989b13fd6-operator-scripts\") pod \"watcher-db-create-7gt2n\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.854029 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-7gt2n"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.942486 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-tbdf5"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.943023 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6jwg\" (UniqueName: \"kubernetes.io/projected/68413519-afb4-4f45-97e0-54b989b13fd6-kube-api-access-v6jwg\") pod \"watcher-db-create-7gt2n\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.943163 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68413519-afb4-4f45-97e0-54b989b13fd6-operator-scripts\") pod \"watcher-db-create-7gt2n\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.943788 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.943881 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68413519-afb4-4f45-97e0-54b989b13fd6-operator-scripts\") pod \"watcher-db-create-7gt2n\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.946368 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.951696 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-tbdf5"] Dec 05 12:37:26 crc kubenswrapper[4807]: I1205 12:37:26.977381 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6jwg\" (UniqueName: \"kubernetes.io/projected/68413519-afb4-4f45-97e0-54b989b13fd6-kube-api-access-v6jwg\") pod \"watcher-db-create-7gt2n\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.146735 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-operator-scripts\") pod \"watcher-test-account-create-update-tbdf5\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.146809 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s98g\" (UniqueName: \"kubernetes.io/projected/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-kube-api-access-5s98g\") pod \"watcher-test-account-create-update-tbdf5\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.154394 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.248275 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3997f615-34be-4ff6-96ce-bf70ed5c2248" path="/var/lib/kubelet/pods/3997f615-34be-4ff6-96ce-bf70ed5c2248/volumes" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.248913 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51f334cd-81eb-447e-a00f-147c84da6027" path="/var/lib/kubelet/pods/51f334cd-81eb-447e-a00f-147c84da6027/volumes" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.249557 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="744b33e1-93da-433c-9abf-319ebc9c5e2e" path="/var/lib/kubelet/pods/744b33e1-93da-433c-9abf-319ebc9c5e2e/volumes" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.251136 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7538542-0f1d-439b-ad97-dd795b74a2c8" path="/var/lib/kubelet/pods/a7538542-0f1d-439b-ad97-dd795b74a2c8/volumes" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.252031 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f11e9916-7522-454b-9341-9bc202abde3b" path="/var/lib/kubelet/pods/f11e9916-7522-454b-9341-9bc202abde3b/volumes" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.252512 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-operator-scripts\") pod \"watcher-test-account-create-update-tbdf5\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.252727 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s98g\" (UniqueName: \"kubernetes.io/projected/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-kube-api-access-5s98g\") pod \"watcher-test-account-create-update-tbdf5\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.259309 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-operator-scripts\") pod \"watcher-test-account-create-update-tbdf5\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.272838 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s98g\" (UniqueName: \"kubernetes.io/projected/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-kube-api-access-5s98g\") pod \"watcher-test-account-create-update-tbdf5\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.558474 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:27 crc kubenswrapper[4807]: I1205 12:37:27.763884 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-7gt2n"] Dec 05 12:37:27 crc kubenswrapper[4807]: W1205 12:37:27.775484 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68413519_afb4_4f45_97e0_54b989b13fd6.slice/crio-1d854c74b061b2469a907e8178244b2104287386855a9d9bf69bfb944ccefc08 WatchSource:0}: Error finding container 1d854c74b061b2469a907e8178244b2104287386855a9d9bf69bfb944ccefc08: Status 404 returned error can't find the container with id 1d854c74b061b2469a907e8178244b2104287386855a9d9bf69bfb944ccefc08 Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.090879 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-tbdf5"] Dec 05 12:37:28 crc kubenswrapper[4807]: W1205 12:37:28.149400 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddac84884_07d6_4bc5_aa0d_a7f1dd579a93.slice/crio-172eb2131a25d446f87eaabb66a41d4e8692ed3c958f215318e869daecfd833a WatchSource:0}: Error finding container 172eb2131a25d446f87eaabb66a41d4e8692ed3c958f215318e869daecfd833a: Status 404 returned error can't find the container with id 172eb2131a25d446f87eaabb66a41d4e8692ed3c958f215318e869daecfd833a Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.235401 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:37:28 crc kubenswrapper[4807]: E1205 12:37:28.235648 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.456361 4807 generic.go:334] "Generic (PLEG): container finished" podID="68413519-afb4-4f45-97e0-54b989b13fd6" containerID="d030f02cd0f90cf768c4d2564f212dd593850c95817efe9c458466124177ca37" exitCode=0 Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.456428 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-7gt2n" event={"ID":"68413519-afb4-4f45-97e0-54b989b13fd6","Type":"ContainerDied","Data":"d030f02cd0f90cf768c4d2564f212dd593850c95817efe9c458466124177ca37"} Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.456457 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-7gt2n" event={"ID":"68413519-afb4-4f45-97e0-54b989b13fd6","Type":"ContainerStarted","Data":"1d854c74b061b2469a907e8178244b2104287386855a9d9bf69bfb944ccefc08"} Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.460289 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerStarted","Data":"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab"} Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.460435 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.460431 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="sg-core" containerID="cri-o://dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" gracePeriod=30 Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.460480 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-notification-agent" containerID="cri-o://4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" gracePeriod=30 Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.460412 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-central-agent" containerID="cri-o://3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" gracePeriod=30 Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.460449 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="proxy-httpd" containerID="cri-o://baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" gracePeriod=30 Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.468899 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" event={"ID":"dac84884-07d6-4bc5-aa0d-a7f1dd579a93","Type":"ContainerStarted","Data":"926d26b4ff019b32dcedffab80d1bf4283c824b401f711f69ea5d40dbd5be1b0"} Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.468956 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" event={"ID":"dac84884-07d6-4bc5-aa0d-a7f1dd579a93","Type":"ContainerStarted","Data":"172eb2131a25d446f87eaabb66a41d4e8692ed3c958f215318e869daecfd833a"} Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.500793 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" podStartSLOduration=2.500764027 podStartE2EDuration="2.500764027s" podCreationTimestamp="2025-12-05 12:37:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:28.494668739 +0000 UTC m=+1877.988532028" watchObservedRunningTime="2025-12-05 12:37:28.500764027 +0000 UTC m=+1877.994627296" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.524270 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.488262901 podStartE2EDuration="6.52425225s" podCreationTimestamp="2025-12-05 12:37:22 +0000 UTC" firstStartedPulling="2025-12-05 12:37:23.199470295 +0000 UTC m=+1872.693333564" lastFinishedPulling="2025-12-05 12:37:27.235459644 +0000 UTC m=+1876.729322913" observedRunningTime="2025-12-05 12:37:28.511710104 +0000 UTC m=+1878.005573383" watchObservedRunningTime="2025-12-05 12:37:28.52425225 +0000 UTC m=+1878.018115519" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.641647 4807 scope.go:117] "RemoveContainer" containerID="a510ff0d47aa643abbacdc68049e138893a0fc057413878ac2825e532904adf4" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.663589 4807 scope.go:117] "RemoveContainer" containerID="c34a04d8ae1613d29f01bc9c4d3b791f4d026f8d7a19a3994535c8cd62445371" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.701462 4807 scope.go:117] "RemoveContainer" containerID="01ab24bdff591ff585f628273fab1e40a730323ec47aa8c84c4a5a3d34ad4343" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.778759 4807 scope.go:117] "RemoveContainer" containerID="09d62b917688f706f0445175c6a63fbd21dfe5b331bba17d470f83240f62aae7" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.807625 4807 scope.go:117] "RemoveContainer" containerID="ce648a9c894f943215c7b189f27414c93f74984371f651e939c8047db0313c24" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.828293 4807 scope.go:117] "RemoveContainer" containerID="95faee3384f11ba7efd6ffe507cf6d168624549dd569d6b97eb36f0fccc62596" Dec 05 12:37:28 crc kubenswrapper[4807]: I1205 12:37:28.850158 4807 scope.go:117] "RemoveContainer" containerID="736c031f3b1359935049cac1b96aca30b54d59751b07a0a4c98167bded987d62" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.034365 4807 scope.go:117] "RemoveContainer" containerID="9ed390e0375a47c630701ff52555a9bc9fba7bb942482a71b08a94bec8bb3a38" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.155881 4807 scope.go:117] "RemoveContainer" containerID="0a6db74cae0d9593cf675f4b80904d6f59d654edce002b35635b186d816e0af5" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.188056 4807 scope.go:117] "RemoveContainer" containerID="ae548c83f5026cc23d83db7dc1f21e86f531b12bc924d0513ec4095641a8c2dc" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.206851 4807 scope.go:117] "RemoveContainer" containerID="e58f694a858405dda6c07db5a11a37916e84824a7b44ad7a5c3f5b327ec6bdf8" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.270271 4807 scope.go:117] "RemoveContainer" containerID="04b61f30585914600acb0b1bf3904adfc07b668dbedabec88fcc9a1c7fd0144f" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.462305 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.487704 4807 generic.go:334] "Generic (PLEG): container finished" podID="dac84884-07d6-4bc5-aa0d-a7f1dd579a93" containerID="926d26b4ff019b32dcedffab80d1bf4283c824b401f711f69ea5d40dbd5be1b0" exitCode=0 Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.487779 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" event={"ID":"dac84884-07d6-4bc5-aa0d-a7f1dd579a93","Type":"ContainerDied","Data":"926d26b4ff019b32dcedffab80d1bf4283c824b401f711f69ea5d40dbd5be1b0"} Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504758 4807 generic.go:334] "Generic (PLEG): container finished" podID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerID="baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" exitCode=0 Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504801 4807 generic.go:334] "Generic (PLEG): container finished" podID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerID="dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" exitCode=2 Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504812 4807 generic.go:334] "Generic (PLEG): container finished" podID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerID="4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" exitCode=0 Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504820 4807 generic.go:334] "Generic (PLEG): container finished" podID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerID="3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" exitCode=0 Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504896 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerDied","Data":"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab"} Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504930 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerDied","Data":"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b"} Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504942 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerDied","Data":"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48"} Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504954 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerDied","Data":"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d"} Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504966 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2f7edd43-2642-40a1-8fc9-70ee2daf650f","Type":"ContainerDied","Data":"b4abaead45f910316db7dadd6bc9ea2b9b9e8d80cb3d9d5023d7558721c296a7"} Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.504983 4807 scope.go:117] "RemoveContainer" containerID="baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.506269 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516114 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-config-data\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516175 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-sg-core-conf-yaml\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516205 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-log-httpd\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516233 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ngxg\" (UniqueName: \"kubernetes.io/projected/2f7edd43-2642-40a1-8fc9-70ee2daf650f-kube-api-access-9ngxg\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516279 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-combined-ca-bundle\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516316 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-run-httpd\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516386 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-scripts\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.516456 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-ceilometer-tls-certs\") pod \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\" (UID: \"2f7edd43-2642-40a1-8fc9-70ee2daf650f\") " Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.530881 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.530917 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f7edd43-2642-40a1-8fc9-70ee2daf650f-kube-api-access-9ngxg" (OuterVolumeSpecName: "kube-api-access-9ngxg") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "kube-api-access-9ngxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.533036 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.582781 4807 scope.go:117] "RemoveContainer" containerID="dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.629637 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-scripts" (OuterVolumeSpecName: "scripts") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.643740 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.643761 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ngxg\" (UniqueName: \"kubernetes.io/projected/2f7edd43-2642-40a1-8fc9-70ee2daf650f-kube-api-access-9ngxg\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.643771 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f7edd43-2642-40a1-8fc9-70ee2daf650f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.643779 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.648598 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.656739 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.657498 4807 scope.go:117] "RemoveContainer" containerID="4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.669628 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-config-data" (OuterVolumeSpecName: "config-data") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.678063 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f7edd43-2642-40a1-8fc9-70ee2daf650f" (UID: "2f7edd43-2642-40a1-8fc9-70ee2daf650f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.678810 4807 scope.go:117] "RemoveContainer" containerID="3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.720640 4807 scope.go:117] "RemoveContainer" containerID="baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.721395 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": container with ID starting with baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab not found: ID does not exist" containerID="baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.721427 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab"} err="failed to get container status \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": rpc error: code = NotFound desc = could not find container \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": container with ID starting with baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.721448 4807 scope.go:117] "RemoveContainer" containerID="dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.721717 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": container with ID starting with dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b not found: ID does not exist" containerID="dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.721736 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b"} err="failed to get container status \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": rpc error: code = NotFound desc = could not find container \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": container with ID starting with dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.721749 4807 scope.go:117] "RemoveContainer" containerID="4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.722004 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": container with ID starting with 4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48 not found: ID does not exist" containerID="4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.722024 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48"} err="failed to get container status \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": rpc error: code = NotFound desc = could not find container \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": container with ID starting with 4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48 not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.722037 4807 scope.go:117] "RemoveContainer" containerID="3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.722203 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": container with ID starting with 3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d not found: ID does not exist" containerID="3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.722224 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d"} err="failed to get container status \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": rpc error: code = NotFound desc = could not find container \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": container with ID starting with 3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.722237 4807 scope.go:117] "RemoveContainer" containerID="baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.722609 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab"} err="failed to get container status \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": rpc error: code = NotFound desc = could not find container \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": container with ID starting with baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.722630 4807 scope.go:117] "RemoveContainer" containerID="dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.722984 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b"} err="failed to get container status \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": rpc error: code = NotFound desc = could not find container \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": container with ID starting with dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723004 4807 scope.go:117] "RemoveContainer" containerID="4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723235 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48"} err="failed to get container status \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": rpc error: code = NotFound desc = could not find container \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": container with ID starting with 4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48 not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723253 4807 scope.go:117] "RemoveContainer" containerID="3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723454 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d"} err="failed to get container status \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": rpc error: code = NotFound desc = could not find container \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": container with ID starting with 3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723473 4807 scope.go:117] "RemoveContainer" containerID="baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723653 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab"} err="failed to get container status \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": rpc error: code = NotFound desc = could not find container \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": container with ID starting with baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723670 4807 scope.go:117] "RemoveContainer" containerID="dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723877 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b"} err="failed to get container status \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": rpc error: code = NotFound desc = could not find container \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": container with ID starting with dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.723897 4807 scope.go:117] "RemoveContainer" containerID="4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.724118 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48"} err="failed to get container status \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": rpc error: code = NotFound desc = could not find container \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": container with ID starting with 4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48 not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.724204 4807 scope.go:117] "RemoveContainer" containerID="3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.724466 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d"} err="failed to get container status \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": rpc error: code = NotFound desc = could not find container \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": container with ID starting with 3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.724555 4807 scope.go:117] "RemoveContainer" containerID="baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.724797 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab"} err="failed to get container status \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": rpc error: code = NotFound desc = could not find container \"baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab\": container with ID starting with baec2458ee3cc7110e7d47b4c1ee6e7d3be09a672495496433c5f8a242298cab not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.724872 4807 scope.go:117] "RemoveContainer" containerID="dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.725810 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b"} err="failed to get container status \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": rpc error: code = NotFound desc = could not find container \"dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b\": container with ID starting with dc96f47f93bea64eb872b724881dbd830e69630ca78b6ce13f8fa226745ef50b not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.725894 4807 scope.go:117] "RemoveContainer" containerID="4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.726189 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48"} err="failed to get container status \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": rpc error: code = NotFound desc = could not find container \"4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48\": container with ID starting with 4b0bb280fc35f7c5d4aee93bac48a4af98553744f21e089803683a4f1f916d48 not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.726258 4807 scope.go:117] "RemoveContainer" containerID="3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.726602 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d"} err="failed to get container status \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": rpc error: code = NotFound desc = could not find container \"3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d\": container with ID starting with 3afee5406d355baa88354b985b66bc3be4543229192e9ec858a927fa07abc43d not found: ID does not exist" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.750843 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.751410 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.751426 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.751442 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7edd43-2642-40a1-8fc9-70ee2daf650f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.856879 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.868164 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883196 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.883651 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="proxy-httpd" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883672 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="proxy-httpd" Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.883713 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-central-agent" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883721 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-central-agent" Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.883738 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-notification-agent" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883744 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-notification-agent" Dec 05 12:37:29 crc kubenswrapper[4807]: E1205 12:37:29.883752 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="sg-core" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883757 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="sg-core" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883941 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-central-agent" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883967 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="proxy-httpd" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883978 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="ceilometer-notification-agent" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.883995 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" containerName="sg-core" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.885493 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.890301 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.890306 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.891404 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.916554 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.933054 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.956464 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-log-httpd\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.956846 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-config-data\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.956944 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5k67\" (UniqueName: \"kubernetes.io/projected/1577bc06-3486-4a3b-adce-2baf9b4eff49-kube-api-access-m5k67\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.957090 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-scripts\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.957193 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.957260 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.957332 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-run-httpd\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:29 crc kubenswrapper[4807]: I1205 12:37:29.957408 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.058354 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6jwg\" (UniqueName: \"kubernetes.io/projected/68413519-afb4-4f45-97e0-54b989b13fd6-kube-api-access-v6jwg\") pod \"68413519-afb4-4f45-97e0-54b989b13fd6\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.058844 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68413519-afb4-4f45-97e0-54b989b13fd6-operator-scripts\") pod \"68413519-afb4-4f45-97e0-54b989b13fd6\" (UID: \"68413519-afb4-4f45-97e0-54b989b13fd6\") " Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059084 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-run-httpd\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059113 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059239 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-log-httpd\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059276 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-config-data\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059311 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5k67\" (UniqueName: \"kubernetes.io/projected/1577bc06-3486-4a3b-adce-2baf9b4eff49-kube-api-access-m5k67\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059356 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-scripts\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059398 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059419 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059463 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68413519-afb4-4f45-97e0-54b989b13fd6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68413519-afb4-4f45-97e0-54b989b13fd6" (UID: "68413519-afb4-4f45-97e0-54b989b13fd6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059726 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-run-httpd\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.059887 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-log-httpd\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.064596 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.068257 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.068436 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-scripts\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.068837 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-config-data\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.074992 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.075666 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68413519-afb4-4f45-97e0-54b989b13fd6-kube-api-access-v6jwg" (OuterVolumeSpecName: "kube-api-access-v6jwg") pod "68413519-afb4-4f45-97e0-54b989b13fd6" (UID: "68413519-afb4-4f45-97e0-54b989b13fd6"). InnerVolumeSpecName "kube-api-access-v6jwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.079432 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5k67\" (UniqueName: \"kubernetes.io/projected/1577bc06-3486-4a3b-adce-2baf9b4eff49-kube-api-access-m5k67\") pod \"ceilometer-0\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.161511 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68413519-afb4-4f45-97e0-54b989b13fd6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.161582 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6jwg\" (UniqueName: \"kubernetes.io/projected/68413519-afb4-4f45-97e0-54b989b13fd6-kube-api-access-v6jwg\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.229551 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.595366 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-7gt2n" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.595366 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-7gt2n" event={"ID":"68413519-afb4-4f45-97e0-54b989b13fd6","Type":"ContainerDied","Data":"1d854c74b061b2469a907e8178244b2104287386855a9d9bf69bfb944ccefc08"} Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.595441 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d854c74b061b2469a907e8178244b2104287386855a9d9bf69bfb944ccefc08" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.678033 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:30 crc kubenswrapper[4807]: W1205 12:37:30.695684 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1577bc06_3486_4a3b_adce_2baf9b4eff49.slice/crio-954a1f67d2fcf7dce3bbe292829ae6d41de1d003c665a4406d11d8117b2288fe WatchSource:0}: Error finding container 954a1f67d2fcf7dce3bbe292829ae6d41de1d003c665a4406d11d8117b2288fe: Status 404 returned error can't find the container with id 954a1f67d2fcf7dce3bbe292829ae6d41de1d003c665a4406d11d8117b2288fe Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.902221 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.973036 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-operator-scripts\") pod \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.973154 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s98g\" (UniqueName: \"kubernetes.io/projected/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-kube-api-access-5s98g\") pod \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\" (UID: \"dac84884-07d6-4bc5-aa0d-a7f1dd579a93\") " Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.973870 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dac84884-07d6-4bc5-aa0d-a7f1dd579a93" (UID: "dac84884-07d6-4bc5-aa0d-a7f1dd579a93"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:37:30 crc kubenswrapper[4807]: I1205 12:37:30.978849 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-kube-api-access-5s98g" (OuterVolumeSpecName: "kube-api-access-5s98g") pod "dac84884-07d6-4bc5-aa0d-a7f1dd579a93" (UID: "dac84884-07d6-4bc5-aa0d-a7f1dd579a93"). InnerVolumeSpecName "kube-api-access-5s98g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.075363 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.075725 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s98g\" (UniqueName: \"kubernetes.io/projected/dac84884-07d6-4bc5-aa0d-a7f1dd579a93-kube-api-access-5s98g\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.250265 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f7edd43-2642-40a1-8fc9-70ee2daf650f" path="/var/lib/kubelet/pods/2f7edd43-2642-40a1-8fc9-70ee2daf650f/volumes" Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.605662 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerStarted","Data":"813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe"} Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.606005 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerStarted","Data":"954a1f67d2fcf7dce3bbe292829ae6d41de1d003c665a4406d11d8117b2288fe"} Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.607456 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" event={"ID":"dac84884-07d6-4bc5-aa0d-a7f1dd579a93","Type":"ContainerDied","Data":"172eb2131a25d446f87eaabb66a41d4e8692ed3c958f215318e869daecfd833a"} Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.607496 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="172eb2131a25d446f87eaabb66a41d4e8692ed3c958f215318e869daecfd833a" Dec 05 12:37:31 crc kubenswrapper[4807]: I1205 12:37:31.607513 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-tbdf5" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.127816 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-c2zws"] Dec 05 12:37:32 crc kubenswrapper[4807]: E1205 12:37:32.128129 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68413519-afb4-4f45-97e0-54b989b13fd6" containerName="mariadb-database-create" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.128142 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="68413519-afb4-4f45-97e0-54b989b13fd6" containerName="mariadb-database-create" Dec 05 12:37:32 crc kubenswrapper[4807]: E1205 12:37:32.128165 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac84884-07d6-4bc5-aa0d-a7f1dd579a93" containerName="mariadb-account-create-update" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.128171 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac84884-07d6-4bc5-aa0d-a7f1dd579a93" containerName="mariadb-account-create-update" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.128317 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="dac84884-07d6-4bc5-aa0d-a7f1dd579a93" containerName="mariadb-account-create-update" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.128338 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="68413519-afb4-4f45-97e0-54b989b13fd6" containerName="mariadb-database-create" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.128861 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.130915 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.130996 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-pkc7z" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.140279 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-c2zws"] Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.192883 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-config-data\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.193256 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcmnm\" (UniqueName: \"kubernetes.io/projected/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-kube-api-access-wcmnm\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.193506 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.193709 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.295328 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-config-data\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.295405 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcmnm\" (UniqueName: \"kubernetes.io/projected/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-kube-api-access-wcmnm\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.295476 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.296415 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.301017 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.301112 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-config-data\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.308028 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.312965 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcmnm\" (UniqueName: \"kubernetes.io/projected/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-kube-api-access-wcmnm\") pod \"watcher-kuttl-db-sync-c2zws\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.443940 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.620851 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerStarted","Data":"4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6"} Dec 05 12:37:32 crc kubenswrapper[4807]: I1205 12:37:32.940224 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-c2zws"] Dec 05 12:37:33 crc kubenswrapper[4807]: I1205 12:37:33.632860 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerStarted","Data":"af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d"} Dec 05 12:37:33 crc kubenswrapper[4807]: I1205 12:37:33.636008 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" event={"ID":"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2","Type":"ContainerStarted","Data":"58097ddc93f4c6390ec8aaaef702269da526e77f5cf41f930789e253b18cc765"} Dec 05 12:37:33 crc kubenswrapper[4807]: I1205 12:37:33.636036 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" event={"ID":"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2","Type":"ContainerStarted","Data":"fb97d1d93a2ecf54b4b545fde839c6cf39128d0b925b501a5101c758cca0fd58"} Dec 05 12:37:33 crc kubenswrapper[4807]: I1205 12:37:33.657926 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" podStartSLOduration=1.6579105360000002 podStartE2EDuration="1.657910536s" podCreationTimestamp="2025-12-05 12:37:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:33.652603968 +0000 UTC m=+1883.146467247" watchObservedRunningTime="2025-12-05 12:37:33.657910536 +0000 UTC m=+1883.151773805" Dec 05 12:37:35 crc kubenswrapper[4807]: I1205 12:37:35.654278 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerStarted","Data":"9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab"} Dec 05 12:37:35 crc kubenswrapper[4807]: I1205 12:37:35.654688 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:35 crc kubenswrapper[4807]: I1205 12:37:35.683326 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.738056148 podStartE2EDuration="6.683302787s" podCreationTimestamp="2025-12-05 12:37:29 +0000 UTC" firstStartedPulling="2025-12-05 12:37:30.698685433 +0000 UTC m=+1880.192548702" lastFinishedPulling="2025-12-05 12:37:34.643932072 +0000 UTC m=+1884.137795341" observedRunningTime="2025-12-05 12:37:35.680086519 +0000 UTC m=+1885.173949788" watchObservedRunningTime="2025-12-05 12:37:35.683302787 +0000 UTC m=+1885.177166056" Dec 05 12:37:36 crc kubenswrapper[4807]: I1205 12:37:36.663269 4807 generic.go:334] "Generic (PLEG): container finished" podID="f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" containerID="58097ddc93f4c6390ec8aaaef702269da526e77f5cf41f930789e253b18cc765" exitCode=0 Dec 05 12:37:36 crc kubenswrapper[4807]: I1205 12:37:36.663376 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" event={"ID":"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2","Type":"ContainerDied","Data":"58097ddc93f4c6390ec8aaaef702269da526e77f5cf41f930789e253b18cc765"} Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.014029 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.103256 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcmnm\" (UniqueName: \"kubernetes.io/projected/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-kube-api-access-wcmnm\") pod \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.103452 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-db-sync-config-data\") pod \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.103537 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-config-data\") pod \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.103603 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-combined-ca-bundle\") pod \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\" (UID: \"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2\") " Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.109610 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-kube-api-access-wcmnm" (OuterVolumeSpecName: "kube-api-access-wcmnm") pod "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" (UID: "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2"). InnerVolumeSpecName "kube-api-access-wcmnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.121451 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" (UID: "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.128421 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" (UID: "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.151920 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-config-data" (OuterVolumeSpecName: "config-data") pod "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" (UID: "f5d2bed9-ec1d-4259-8b58-a01ae6909fc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.205222 4807 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.205263 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.205278 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.205289 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcmnm\" (UniqueName: \"kubernetes.io/projected/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2-kube-api-access-wcmnm\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.685914 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" event={"ID":"f5d2bed9-ec1d-4259-8b58-a01ae6909fc2","Type":"ContainerDied","Data":"fb97d1d93a2ecf54b4b545fde839c6cf39128d0b925b501a5101c758cca0fd58"} Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.685959 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb97d1d93a2ecf54b4b545fde839c6cf39128d0b925b501a5101c758cca0fd58" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.686039 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-c2zws" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.973898 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:38 crc kubenswrapper[4807]: E1205 12:37:38.974230 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" containerName="watcher-kuttl-db-sync" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.974252 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" containerName="watcher-kuttl-db-sync" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.974416 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" containerName="watcher-kuttl-db-sync" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.975002 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.977710 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-pkc7z" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.977945 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.989590 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.991256 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.993618 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Dec 05 12:37:38 crc kubenswrapper[4807]: I1205 12:37:38.999850 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.033989 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.048218 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.048375 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.061506 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.062621 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.066325 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.067047 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082212 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082261 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvcsv\" (UniqueName: \"kubernetes.io/projected/e849db0d-4bb4-48aa-bb8f-8730d424694b-kube-api-access-nvcsv\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082314 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082341 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a51fac-9422-488c-922a-c42e21ccd7df-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082364 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082383 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e849db0d-4bb4-48aa-bb8f-8730d424694b-logs\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082406 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082424 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kch9r\" (UniqueName: \"kubernetes.io/projected/f1a51fac-9422-488c-922a-c42e21ccd7df-kube-api-access-kch9r\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082446 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082469 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88b4b699-36f9-4259-9fb6-f241f12550c9-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082491 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-logs\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082546 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082577 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082619 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082675 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082729 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082758 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082789 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082826 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082853 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5n5b\" (UniqueName: \"kubernetes.io/projected/88b4b699-36f9-4259-9fb6-f241f12550c9-kube-api-access-g5n5b\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082880 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082914 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27rf7\" (UniqueName: \"kubernetes.io/projected/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-kube-api-access-27rf7\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.082941 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.098358 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184076 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184138 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a51fac-9422-488c-922a-c42e21ccd7df-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184167 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184187 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e849db0d-4bb4-48aa-bb8f-8730d424694b-logs\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184208 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184230 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kch9r\" (UniqueName: \"kubernetes.io/projected/f1a51fac-9422-488c-922a-c42e21ccd7df-kube-api-access-kch9r\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184254 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184275 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88b4b699-36f9-4259-9fb6-f241f12550c9-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184296 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-logs\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184331 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184360 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184394 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184442 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184467 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184494 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184540 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184573 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184596 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5n5b\" (UniqueName: \"kubernetes.io/projected/88b4b699-36f9-4259-9fb6-f241f12550c9-kube-api-access-g5n5b\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184621 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184652 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27rf7\" (UniqueName: \"kubernetes.io/projected/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-kube-api-access-27rf7\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184678 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184707 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184727 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e849db0d-4bb4-48aa-bb8f-8730d424694b-logs\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184728 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvcsv\" (UniqueName: \"kubernetes.io/projected/e849db0d-4bb4-48aa-bb8f-8730d424694b-kube-api-access-nvcsv\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184780 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88b4b699-36f9-4259-9fb6-f241f12550c9-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.184813 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-logs\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.185394 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a51fac-9422-488c-922a-c42e21ccd7df-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.193314 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.194288 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.199553 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.199994 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.200027 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.200000 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.200094 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.200485 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.200759 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.203144 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.209555 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvcsv\" (UniqueName: \"kubernetes.io/projected/e849db0d-4bb4-48aa-bb8f-8730d424694b-kube-api-access-nvcsv\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.209867 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.210377 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.211100 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.211692 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.224069 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5n5b\" (UniqueName: \"kubernetes.io/projected/88b4b699-36f9-4259-9fb6-f241f12550c9-kube-api-access-g5n5b\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.232138 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kch9r\" (UniqueName: \"kubernetes.io/projected/f1a51fac-9422-488c-922a-c42e21ccd7df-kube-api-access-kch9r\") pod \"watcher-kuttl-applier-0\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.239356 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27rf7\" (UniqueName: \"kubernetes.io/projected/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-kube-api-access-27rf7\") pod \"watcher-kuttl-api-1\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.239408 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:37:39 crc kubenswrapper[4807]: E1205 12:37:39.239667 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.242290 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.317934 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.338880 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.388975 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:39 crc kubenswrapper[4807]: I1205 12:37:39.403457 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.042216 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.120967 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.160844 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:37:40 crc kubenswrapper[4807]: W1205 12:37:40.163642 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode849db0d_4bb4_48aa_bb8f_8730d424694b.slice/crio-7b94f60d30cb812201e81c535efcad12657bd68c96089c4f94d1a7a832b7c84d WatchSource:0}: Error finding container 7b94f60d30cb812201e81c535efcad12657bd68c96089c4f94d1a7a832b7c84d: Status 404 returned error can't find the container with id 7b94f60d30cb812201e81c535efcad12657bd68c96089c4f94d1a7a832b7c84d Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.191498 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.731717 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7","Type":"ContainerStarted","Data":"2d4337a6b57e2021986af3b857969d76dc8a42c6797f3452d00a49dfeaf5830d"} Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.734997 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"88b4b699-36f9-4259-9fb6-f241f12550c9","Type":"ContainerStarted","Data":"5dc7ede7ced29bc29a5d9c917f2857211d9e8ea15deaede18c619646eb45e563"} Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.740219 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"f1a51fac-9422-488c-922a-c42e21ccd7df","Type":"ContainerStarted","Data":"45ae363cd449d04b185adaef691d161424829e7d1ad2f3506ea5cdcd3f652e23"} Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.742215 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"e849db0d-4bb4-48aa-bb8f-8730d424694b","Type":"ContainerStarted","Data":"7b94f60d30cb812201e81c535efcad12657bd68c96089c4f94d1a7a832b7c84d"} Dec 05 12:37:40 crc kubenswrapper[4807]: I1205 12:37:40.754473 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.75445948 podStartE2EDuration="2.75445948s" podCreationTimestamp="2025-12-05 12:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:40.750211937 +0000 UTC m=+1890.244075206" watchObservedRunningTime="2025-12-05 12:37:40.75445948 +0000 UTC m=+1890.248322749" Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.262690 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=3.262666818 podStartE2EDuration="3.262666818s" podCreationTimestamp="2025-12-05 12:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:40.78809088 +0000 UTC m=+1890.281954149" watchObservedRunningTime="2025-12-05 12:37:41.262666818 +0000 UTC m=+1890.756530097" Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.751211 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"88b4b699-36f9-4259-9fb6-f241f12550c9","Type":"ContainerStarted","Data":"2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9"} Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.752806 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"f1a51fac-9422-488c-922a-c42e21ccd7df","Type":"ContainerStarted","Data":"4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e"} Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.754324 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"e849db0d-4bb4-48aa-bb8f-8730d424694b","Type":"ContainerStarted","Data":"e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1"} Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.754357 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"e849db0d-4bb4-48aa-bb8f-8730d424694b","Type":"ContainerStarted","Data":"75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c"} Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.754689 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.756395 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7","Type":"ContainerStarted","Data":"2338f381c9e544e5774aaa38220b2ef7ce00304710f06439b086c7159712abaf"} Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.756439 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7","Type":"ContainerStarted","Data":"32958eadddc2a80e25cbd731c2470867f6b5e87c8751430309994021eb0bfa48"} Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.756656 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.776670 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=3.776651487 podStartE2EDuration="3.776651487s" podCreationTimestamp="2025-12-05 12:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:41.773218673 +0000 UTC m=+1891.267081942" watchObservedRunningTime="2025-12-05 12:37:41.776651487 +0000 UTC m=+1891.270514756" Dec 05 12:37:41 crc kubenswrapper[4807]: I1205 12:37:41.792833 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-1" podStartSLOduration=3.7928178900000002 podStartE2EDuration="3.79281789s" podCreationTimestamp="2025-12-05 12:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:41.78950477 +0000 UTC m=+1891.283368049" watchObservedRunningTime="2025-12-05 12:37:41.79281789 +0000 UTC m=+1891.286681159" Dec 05 12:37:44 crc kubenswrapper[4807]: I1205 12:37:44.319024 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:44 crc kubenswrapper[4807]: I1205 12:37:44.330380 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:44 crc kubenswrapper[4807]: I1205 12:37:44.339900 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:44 crc kubenswrapper[4807]: I1205 12:37:44.340239 4807 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:37:44 crc kubenswrapper[4807]: I1205 12:37:44.377216 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:44 crc kubenswrapper[4807]: I1205 12:37:44.389353 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.318843 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.339690 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.345077 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.347253 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.389579 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.397664 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.404658 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.441806 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.822895 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.827635 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.830061 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.856798 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:37:49 crc kubenswrapper[4807]: I1205 12:37:49.869175 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:37:51 crc kubenswrapper[4807]: I1205 12:37:51.855043 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:51 crc kubenswrapper[4807]: I1205 12:37:51.855661 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-central-agent" containerID="cri-o://813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe" gracePeriod=30 Dec 05 12:37:51 crc kubenswrapper[4807]: I1205 12:37:51.855789 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-notification-agent" containerID="cri-o://4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6" gracePeriod=30 Dec 05 12:37:51 crc kubenswrapper[4807]: I1205 12:37:51.855755 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="sg-core" containerID="cri-o://af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d" gracePeriod=30 Dec 05 12:37:51 crc kubenswrapper[4807]: I1205 12:37:51.855906 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="proxy-httpd" containerID="cri-o://9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab" gracePeriod=30 Dec 05 12:37:51 crc kubenswrapper[4807]: I1205 12:37:51.892640 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.218:3000/\": read tcp 10.217.0.2:43346->10.217.0.218:3000: read: connection reset by peer" Dec 05 12:37:52 crc kubenswrapper[4807]: I1205 12:37:52.235341 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:37:52 crc kubenswrapper[4807]: E1205 12:37:52.235668 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:37:52 crc kubenswrapper[4807]: I1205 12:37:52.850794 4807 generic.go:334] "Generic (PLEG): container finished" podID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerID="9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab" exitCode=0 Dec 05 12:37:52 crc kubenswrapper[4807]: I1205 12:37:52.851102 4807 generic.go:334] "Generic (PLEG): container finished" podID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerID="af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d" exitCode=2 Dec 05 12:37:52 crc kubenswrapper[4807]: I1205 12:37:52.851112 4807 generic.go:334] "Generic (PLEG): container finished" podID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerID="813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe" exitCode=0 Dec 05 12:37:52 crc kubenswrapper[4807]: I1205 12:37:52.850864 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerDied","Data":"9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab"} Dec 05 12:37:52 crc kubenswrapper[4807]: I1205 12:37:52.851145 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerDied","Data":"af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d"} Dec 05 12:37:52 crc kubenswrapper[4807]: I1205 12:37:52.851161 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerDied","Data":"813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe"} Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.695778 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.852733 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-config-data\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.852972 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-ceilometer-tls-certs\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853003 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-combined-ca-bundle\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853022 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-scripts\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853061 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-sg-core-conf-yaml\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853126 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-log-httpd\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853240 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-run-httpd\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853377 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5k67\" (UniqueName: \"kubernetes.io/projected/1577bc06-3486-4a3b-adce-2baf9b4eff49-kube-api-access-m5k67\") pod \"1577bc06-3486-4a3b-adce-2baf9b4eff49\" (UID: \"1577bc06-3486-4a3b-adce-2baf9b4eff49\") " Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853923 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.853972 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.866929 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1577bc06-3486-4a3b-adce-2baf9b4eff49-kube-api-access-m5k67" (OuterVolumeSpecName: "kube-api-access-m5k67") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "kube-api-access-m5k67". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.867726 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-scripts" (OuterVolumeSpecName: "scripts") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.873702 4807 generic.go:334] "Generic (PLEG): container finished" podID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerID="4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6" exitCode=0 Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.873743 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerDied","Data":"4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6"} Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.873777 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1577bc06-3486-4a3b-adce-2baf9b4eff49","Type":"ContainerDied","Data":"954a1f67d2fcf7dce3bbe292829ae6d41de1d003c665a4406d11d8117b2288fe"} Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.873798 4807 scope.go:117] "RemoveContainer" containerID="9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.873817 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.878934 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.912024 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.922005 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.959165 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.959217 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.959229 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.959241 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.959252 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.959267 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1577bc06-3486-4a3b-adce-2baf9b4eff49-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.959351 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5k67\" (UniqueName: \"kubernetes.io/projected/1577bc06-3486-4a3b-adce-2baf9b4eff49-kube-api-access-m5k67\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.964945 4807 scope.go:117] "RemoveContainer" containerID="af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.967921 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-config-data" (OuterVolumeSpecName: "config-data") pod "1577bc06-3486-4a3b-adce-2baf9b4eff49" (UID: "1577bc06-3486-4a3b-adce-2baf9b4eff49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:54 crc kubenswrapper[4807]: I1205 12:37:54.985166 4807 scope.go:117] "RemoveContainer" containerID="4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.002042 4807 scope.go:117] "RemoveContainer" containerID="813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.021073 4807 scope.go:117] "RemoveContainer" containerID="9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab" Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.021479 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab\": container with ID starting with 9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab not found: ID does not exist" containerID="9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.021512 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab"} err="failed to get container status \"9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab\": rpc error: code = NotFound desc = could not find container \"9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab\": container with ID starting with 9994753c24f03e97a2f9dced6d4dfbb39296c3599e4dabdc6215b11565f092ab not found: ID does not exist" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.021553 4807 scope.go:117] "RemoveContainer" containerID="af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d" Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.021843 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d\": container with ID starting with af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d not found: ID does not exist" containerID="af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.021967 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d"} err="failed to get container status \"af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d\": rpc error: code = NotFound desc = could not find container \"af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d\": container with ID starting with af087cae4bcbee947fae1a13776e497e5521f3eb72cacb20d227c8ff09916d5d not found: ID does not exist" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.022067 4807 scope.go:117] "RemoveContainer" containerID="4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6" Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.022456 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6\": container with ID starting with 4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6 not found: ID does not exist" containerID="4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.022479 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6"} err="failed to get container status \"4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6\": rpc error: code = NotFound desc = could not find container \"4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6\": container with ID starting with 4bb920a10dd0722fd7f84e9e042fe21986c302450206307ec1f216555580a2c6 not found: ID does not exist" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.022495 4807 scope.go:117] "RemoveContainer" containerID="813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe" Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.022787 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe\": container with ID starting with 813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe not found: ID does not exist" containerID="813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.022809 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe"} err="failed to get container status \"813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe\": rpc error: code = NotFound desc = could not find container \"813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe\": container with ID starting with 813c0fb592444f90c79a4090ed2d4f54da7a8e245782f0d3a38d7a03e11c75fe not found: ID does not exist" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.060691 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1577bc06-3486-4a3b-adce-2baf9b4eff49-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.202492 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.213040 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.286998 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" path="/var/lib/kubelet/pods/1577bc06-3486-4a3b-adce-2baf9b4eff49/volumes" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.287722 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.288024 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="sg-core" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288044 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="sg-core" Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.288061 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-notification-agent" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288069 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-notification-agent" Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.288084 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-central-agent" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288092 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-central-agent" Dec 05 12:37:55 crc kubenswrapper[4807]: E1205 12:37:55.288121 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="proxy-httpd" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288129 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="proxy-httpd" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288360 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="sg-core" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288379 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-notification-agent" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288393 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="proxy-httpd" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.288412 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="1577bc06-3486-4a3b-adce-2baf9b4eff49" containerName="ceilometer-central-agent" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.290395 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.293634 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.293916 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.293960 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.296243 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.365812 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.365876 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.365898 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-config-data\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.365922 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.366142 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crzjw\" (UniqueName: \"kubernetes.io/projected/58286ab5-dd9b-463c-95af-5ae2872ed8b4-kube-api-access-crzjw\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.366284 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-log-httpd\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.366383 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-scripts\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.366470 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-run-httpd\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.467642 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-log-httpd\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.467935 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-scripts\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.467974 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-run-httpd\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.468036 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.468083 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.468107 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-config-data\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.468141 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.468189 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crzjw\" (UniqueName: \"kubernetes.io/projected/58286ab5-dd9b-463c-95af-5ae2872ed8b4-kube-api-access-crzjw\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.468406 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-log-httpd\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.469017 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-run-httpd\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.474038 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-scripts\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.474062 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.474730 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.475097 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.475789 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-config-data\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.486999 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crzjw\" (UniqueName: \"kubernetes.io/projected/58286ab5-dd9b-463c-95af-5ae2872ed8b4-kube-api-access-crzjw\") pod \"ceilometer-0\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:55 crc kubenswrapper[4807]: I1205 12:37:55.606761 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:37:56 crc kubenswrapper[4807]: W1205 12:37:56.178643 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58286ab5_dd9b_463c_95af_5ae2872ed8b4.slice/crio-ee8c3dbdb3cc60a8324d721a50515a63c4c18787411acda6d4277323755e0522 WatchSource:0}: Error finding container ee8c3dbdb3cc60a8324d721a50515a63c4c18787411acda6d4277323755e0522: Status 404 returned error can't find the container with id ee8c3dbdb3cc60a8324d721a50515a63c4c18787411acda6d4277323755e0522 Dec 05 12:37:56 crc kubenswrapper[4807]: I1205 12:37:56.179977 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:37:56 crc kubenswrapper[4807]: I1205 12:37:56.903120 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerStarted","Data":"7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5"} Dec 05 12:37:56 crc kubenswrapper[4807]: I1205 12:37:56.903879 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerStarted","Data":"ee8c3dbdb3cc60a8324d721a50515a63c4c18787411acda6d4277323755e0522"} Dec 05 12:37:57 crc kubenswrapper[4807]: I1205 12:37:57.912736 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerStarted","Data":"d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db"} Dec 05 12:37:58 crc kubenswrapper[4807]: I1205 12:37:58.925083 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerStarted","Data":"a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e"} Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.135836 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v"] Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.137027 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.142054 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.142242 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-scripts" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.163699 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v"] Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.250677 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q78qk\" (UniqueName: \"kubernetes.io/projected/e786ea23-5fd8-4328-9d88-8d65b94ffd33-kube-api-access-q78qk\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.250996 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-combined-ca-bundle\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.251169 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-scripts-volume\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.251283 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-config-data\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.353009 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-scripts-volume\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.353074 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-config-data\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.353147 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q78qk\" (UniqueName: \"kubernetes.io/projected/e786ea23-5fd8-4328-9d88-8d65b94ffd33-kube-api-access-q78qk\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.353185 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-combined-ca-bundle\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.358639 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-scripts-volume\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.358739 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-combined-ca-bundle\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.364162 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-config-data\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.383020 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q78qk\" (UniqueName: \"kubernetes.io/projected/e786ea23-5fd8-4328-9d88-8d65b94ffd33-kube-api-access-q78qk\") pod \"watcher-kuttl-db-purge-29415638-xwj4v\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.456622 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.887411 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v"] Dec 05 12:38:00 crc kubenswrapper[4807]: W1205 12:38:00.893873 4807 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode786ea23_5fd8_4328_9d88_8d65b94ffd33.slice/crio-adf5b22c89504e541ad5b7caf4b959de13aac8a57304dd029f615c4211b4efcd WatchSource:0}: Error finding container adf5b22c89504e541ad5b7caf4b959de13aac8a57304dd029f615c4211b4efcd: Status 404 returned error can't find the container with id adf5b22c89504e541ad5b7caf4b959de13aac8a57304dd029f615c4211b4efcd Dec 05 12:38:00 crc kubenswrapper[4807]: I1205 12:38:00.943211 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" event={"ID":"e786ea23-5fd8-4328-9d88-8d65b94ffd33","Type":"ContainerStarted","Data":"adf5b22c89504e541ad5b7caf4b959de13aac8a57304dd029f615c4211b4efcd"} Dec 05 12:38:01 crc kubenswrapper[4807]: I1205 12:38:01.952085 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" event={"ID":"e786ea23-5fd8-4328-9d88-8d65b94ffd33","Type":"ContainerStarted","Data":"ffa53385c003261df2402b6c6df4263734c55aee847f8445af2f89e163c815df"} Dec 05 12:38:01 crc kubenswrapper[4807]: I1205 12:38:01.953997 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerStarted","Data":"ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965"} Dec 05 12:38:01 crc kubenswrapper[4807]: I1205 12:38:01.954126 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:01 crc kubenswrapper[4807]: I1205 12:38:01.978429 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" podStartSLOduration=1.9784093280000001 podStartE2EDuration="1.978409328s" podCreationTimestamp="2025-12-05 12:38:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:38:01.97353982 +0000 UTC m=+1911.467403089" watchObservedRunningTime="2025-12-05 12:38:01.978409328 +0000 UTC m=+1911.472272607" Dec 05 12:38:02 crc kubenswrapper[4807]: I1205 12:38:02.002216 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.368023837 podStartE2EDuration="7.002193738s" podCreationTimestamp="2025-12-05 12:37:55 +0000 UTC" firstStartedPulling="2025-12-05 12:37:56.181197128 +0000 UTC m=+1905.675060397" lastFinishedPulling="2025-12-05 12:38:00.815367029 +0000 UTC m=+1910.309230298" observedRunningTime="2025-12-05 12:38:01.998285593 +0000 UTC m=+1911.492148872" watchObservedRunningTime="2025-12-05 12:38:02.002193738 +0000 UTC m=+1911.496057007" Dec 05 12:38:03 crc kubenswrapper[4807]: I1205 12:38:03.969853 4807 generic.go:334] "Generic (PLEG): container finished" podID="e786ea23-5fd8-4328-9d88-8d65b94ffd33" containerID="ffa53385c003261df2402b6c6df4263734c55aee847f8445af2f89e163c815df" exitCode=0 Dec 05 12:38:03 crc kubenswrapper[4807]: I1205 12:38:03.969947 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" event={"ID":"e786ea23-5fd8-4328-9d88-8d65b94ffd33","Type":"ContainerDied","Data":"ffa53385c003261df2402b6c6df4263734c55aee847f8445af2f89e163c815df"} Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.337262 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.428963 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-combined-ca-bundle\") pod \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.429197 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q78qk\" (UniqueName: \"kubernetes.io/projected/e786ea23-5fd8-4328-9d88-8d65b94ffd33-kube-api-access-q78qk\") pod \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.429229 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-config-data\") pod \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.429258 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-scripts-volume\") pod \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\" (UID: \"e786ea23-5fd8-4328-9d88-8d65b94ffd33\") " Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.439115 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-scripts-volume" (OuterVolumeSpecName: "scripts-volume") pod "e786ea23-5fd8-4328-9d88-8d65b94ffd33" (UID: "e786ea23-5fd8-4328-9d88-8d65b94ffd33"). InnerVolumeSpecName "scripts-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.449850 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e786ea23-5fd8-4328-9d88-8d65b94ffd33-kube-api-access-q78qk" (OuterVolumeSpecName: "kube-api-access-q78qk") pod "e786ea23-5fd8-4328-9d88-8d65b94ffd33" (UID: "e786ea23-5fd8-4328-9d88-8d65b94ffd33"). InnerVolumeSpecName "kube-api-access-q78qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.461237 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e786ea23-5fd8-4328-9d88-8d65b94ffd33" (UID: "e786ea23-5fd8-4328-9d88-8d65b94ffd33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.482167 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-config-data" (OuterVolumeSpecName: "config-data") pod "e786ea23-5fd8-4328-9d88-8d65b94ffd33" (UID: "e786ea23-5fd8-4328-9d88-8d65b94ffd33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.530712 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q78qk\" (UniqueName: \"kubernetes.io/projected/e786ea23-5fd8-4328-9d88-8d65b94ffd33-kube-api-access-q78qk\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.530759 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.530774 4807 reconciler_common.go:293] "Volume detached for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-scripts-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:05 crc kubenswrapper[4807]: I1205 12:38:05.530784 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e786ea23-5fd8-4328-9d88-8d65b94ffd33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:06 crc kubenswrapper[4807]: I1205 12:38:06.008940 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" event={"ID":"e786ea23-5fd8-4328-9d88-8d65b94ffd33","Type":"ContainerDied","Data":"adf5b22c89504e541ad5b7caf4b959de13aac8a57304dd029f615c4211b4efcd"} Dec 05 12:38:06 crc kubenswrapper[4807]: I1205 12:38:06.008985 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adf5b22c89504e541ad5b7caf4b959de13aac8a57304dd029f615c4211b4efcd" Dec 05 12:38:06 crc kubenswrapper[4807]: I1205 12:38:06.009056 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v" Dec 05 12:38:06 crc kubenswrapper[4807]: I1205 12:38:06.235225 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:38:07 crc kubenswrapper[4807]: I1205 12:38:07.026747 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"5d1db79e77bd059c1448ecb5528526ada5577ec3cff58469fb098ce90654d049"} Dec 05 12:38:07 crc kubenswrapper[4807]: I1205 12:38:07.049667 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kmv6v"] Dec 05 12:38:07 crc kubenswrapper[4807]: I1205 12:38:07.061340 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kmv6v"] Dec 05 12:38:07 crc kubenswrapper[4807]: I1205 12:38:07.245677 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c70e7072-33ae-46e7-9cae-7a7c9b9e80a6" path="/var/lib/kubelet/pods/c70e7072-33ae-46e7-9cae-7a7c9b9e80a6/volumes" Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.914982 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-c2zws"] Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.921614 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-c2zws"] Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.933662 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v"] Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.942579 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29415638-xwj4v"] Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.982352 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-sr7cp"] Dec 05 12:38:09 crc kubenswrapper[4807]: E1205 12:38:09.982754 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e786ea23-5fd8-4328-9d88-8d65b94ffd33" containerName="watcher-db-manage" Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.982772 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="e786ea23-5fd8-4328-9d88-8d65b94ffd33" containerName="watcher-db-manage" Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.982926 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="e786ea23-5fd8-4328-9d88-8d65b94ffd33" containerName="watcher-db-manage" Dec 05 12:38:09 crc kubenswrapper[4807]: I1205 12:38:09.983507 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.030032 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-operator-scripts\") pod \"watchertest-account-delete-sr7cp\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.030227 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np8gq\" (UniqueName: \"kubernetes.io/projected/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-kube-api-access-np8gq\") pod \"watchertest-account-delete-sr7cp\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.038226 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-sr7cp"] Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.060154 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.060476 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="f1a51fac-9422-488c-922a-c42e21ccd7df" containerName="watcher-applier" containerID="cri-o://4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e" gracePeriod=30 Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.074721 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.074988 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="88b4b699-36f9-4259-9fb6-f241f12550c9" containerName="watcher-decision-engine" containerID="cri-o://2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9" gracePeriod=30 Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.097394 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.097702 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-kuttl-api-log" containerID="cri-o://e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1" gracePeriod=30 Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.098153 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-api" containerID="cri-o://75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c" gracePeriod=30 Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.126066 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.126333 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-kuttl-api-log" containerID="cri-o://2338f381c9e544e5774aaa38220b2ef7ce00304710f06439b086c7159712abaf" gracePeriod=30 Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.127915 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-api" containerID="cri-o://32958eadddc2a80e25cbd731c2470867f6b5e87c8751430309994021eb0bfa48" gracePeriod=30 Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.131591 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np8gq\" (UniqueName: \"kubernetes.io/projected/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-kube-api-access-np8gq\") pod \"watchertest-account-delete-sr7cp\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.131710 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-operator-scripts\") pod \"watchertest-account-delete-sr7cp\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.132680 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-operator-scripts\") pod \"watchertest-account-delete-sr7cp\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.161714 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np8gq\" (UniqueName: \"kubernetes.io/projected/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-kube-api-access-np8gq\") pod \"watchertest-account-delete-sr7cp\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.345916 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:10 crc kubenswrapper[4807]: I1205 12:38:10.862896 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-sr7cp"] Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.064562 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" event={"ID":"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9","Type":"ContainerStarted","Data":"e8f4077c43553a2d793097a71fd264968c4904bb69df59dc0e0694d93e606976"} Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.066602 4807 generic.go:334] "Generic (PLEG): container finished" podID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerID="e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1" exitCode=143 Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.066669 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"e849db0d-4bb4-48aa-bb8f-8730d424694b","Type":"ContainerDied","Data":"e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1"} Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.068599 4807 generic.go:334] "Generic (PLEG): container finished" podID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerID="2338f381c9e544e5774aaa38220b2ef7ce00304710f06439b086c7159712abaf" exitCode=143 Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.068627 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7","Type":"ContainerDied","Data":"2338f381c9e544e5774aaa38220b2ef7ce00304710f06439b086c7159712abaf"} Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.255984 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e786ea23-5fd8-4328-9d88-8d65b94ffd33" path="/var/lib/kubelet/pods/e786ea23-5fd8-4328-9d88-8d65b94ffd33/volumes" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.257045 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d2bed9-ec1d-4259-8b58-a01ae6909fc2" path="/var/lib/kubelet/pods/f5d2bed9-ec1d-4259-8b58-a01ae6909fc2/volumes" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.575555 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.756202 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-combined-ca-bundle\") pod \"e849db0d-4bb4-48aa-bb8f-8730d424694b\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.756288 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvcsv\" (UniqueName: \"kubernetes.io/projected/e849db0d-4bb4-48aa-bb8f-8730d424694b-kube-api-access-nvcsv\") pod \"e849db0d-4bb4-48aa-bb8f-8730d424694b\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.756319 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e849db0d-4bb4-48aa-bb8f-8730d424694b-logs\") pod \"e849db0d-4bb4-48aa-bb8f-8730d424694b\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.756380 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-cert-memcached-mtls\") pod \"e849db0d-4bb4-48aa-bb8f-8730d424694b\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.757016 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e849db0d-4bb4-48aa-bb8f-8730d424694b-logs" (OuterVolumeSpecName: "logs") pod "e849db0d-4bb4-48aa-bb8f-8730d424694b" (UID: "e849db0d-4bb4-48aa-bb8f-8730d424694b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.757107 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-config-data\") pod \"e849db0d-4bb4-48aa-bb8f-8730d424694b\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.757428 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-custom-prometheus-ca\") pod \"e849db0d-4bb4-48aa-bb8f-8730d424694b\" (UID: \"e849db0d-4bb4-48aa-bb8f-8730d424694b\") " Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.758015 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e849db0d-4bb4-48aa-bb8f-8730d424694b-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.762846 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e849db0d-4bb4-48aa-bb8f-8730d424694b-kube-api-access-nvcsv" (OuterVolumeSpecName: "kube-api-access-nvcsv") pod "e849db0d-4bb4-48aa-bb8f-8730d424694b" (UID: "e849db0d-4bb4-48aa-bb8f-8730d424694b"). InnerVolumeSpecName "kube-api-access-nvcsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.806697 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "e849db0d-4bb4-48aa-bb8f-8730d424694b" (UID: "e849db0d-4bb4-48aa-bb8f-8730d424694b"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.822010 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e849db0d-4bb4-48aa-bb8f-8730d424694b" (UID: "e849db0d-4bb4-48aa-bb8f-8730d424694b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.826131 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-config-data" (OuterVolumeSpecName: "config-data") pod "e849db0d-4bb4-48aa-bb8f-8730d424694b" (UID: "e849db0d-4bb4-48aa-bb8f-8730d424694b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.857734 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "e849db0d-4bb4-48aa-bb8f-8730d424694b" (UID: "e849db0d-4bb4-48aa-bb8f-8730d424694b"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.859734 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.859758 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.859770 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvcsv\" (UniqueName: \"kubernetes.io/projected/e849db0d-4bb4-48aa-bb8f-8730d424694b-kube-api-access-nvcsv\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.859782 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:11 crc kubenswrapper[4807]: I1205 12:38:11.859794 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e849db0d-4bb4-48aa-bb8f-8730d424694b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.079050 4807 generic.go:334] "Generic (PLEG): container finished" podID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerID="75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c" exitCode=0 Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.079121 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.079129 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"e849db0d-4bb4-48aa-bb8f-8730d424694b","Type":"ContainerDied","Data":"75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c"} Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.079515 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"e849db0d-4bb4-48aa-bb8f-8730d424694b","Type":"ContainerDied","Data":"7b94f60d30cb812201e81c535efcad12657bd68c96089c4f94d1a7a832b7c84d"} Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.079553 4807 scope.go:117] "RemoveContainer" containerID="75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.081694 4807 generic.go:334] "Generic (PLEG): container finished" podID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerID="32958eadddc2a80e25cbd731c2470867f6b5e87c8751430309994021eb0bfa48" exitCode=0 Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.081780 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7","Type":"ContainerDied","Data":"32958eadddc2a80e25cbd731c2470867f6b5e87c8751430309994021eb0bfa48"} Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.083034 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" event={"ID":"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9","Type":"ContainerStarted","Data":"fedbebd8ae752c9fa6d9d23f8d3a5b5c3fb6eccda16303307b316631ec932efe"} Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.104763 4807 scope.go:117] "RemoveContainer" containerID="e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.108877 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" podStartSLOduration=3.108849026 podStartE2EDuration="3.108849026s" podCreationTimestamp="2025-12-05 12:38:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:38:12.101328962 +0000 UTC m=+1921.595192231" watchObservedRunningTime="2025-12-05 12:38:12.108849026 +0000 UTC m=+1921.602712305" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.129115 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.129155 4807 scope.go:117] "RemoveContainer" containerID="75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c" Dec 05 12:38:12 crc kubenswrapper[4807]: E1205 12:38:12.129708 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c\": container with ID starting with 75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c not found: ID does not exist" containerID="75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.129761 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c"} err="failed to get container status \"75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c\": rpc error: code = NotFound desc = could not find container \"75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c\": container with ID starting with 75cf8021d17ed3ba1b5f5770b10b0cfe6928f458f5105780af7f548c51ec640c not found: ID does not exist" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.129788 4807 scope.go:117] "RemoveContainer" containerID="e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1" Dec 05 12:38:12 crc kubenswrapper[4807]: E1205 12:38:12.130065 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1\": container with ID starting with e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1 not found: ID does not exist" containerID="e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.130085 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1"} err="failed to get container status \"e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1\": rpc error: code = NotFound desc = could not find container \"e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1\": container with ID starting with e00271b00f8b25250699a3a1a5afa7364150e935e16e00ffdee934045e1707c1 not found: ID does not exist" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.137872 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.638858 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.682822 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-config-data\") pod \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.683215 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-logs\") pod \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.683242 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-custom-prometheus-ca\") pod \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.683352 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27rf7\" (UniqueName: \"kubernetes.io/projected/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-kube-api-access-27rf7\") pod \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.686980 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-cert-memcached-mtls\") pod \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.687012 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-combined-ca-bundle\") pod \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\" (UID: \"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.687463 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-logs" (OuterVolumeSpecName: "logs") pod "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" (UID: "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.687631 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.693163 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-kube-api-access-27rf7" (OuterVolumeSpecName: "kube-api-access-27rf7") pod "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" (UID: "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7"). InnerVolumeSpecName "kube-api-access-27rf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.758645 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" (UID: "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.781121 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" (UID: "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.786357 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.788757 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.788988 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27rf7\" (UniqueName: \"kubernetes.io/projected/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-kube-api-access-27rf7\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.789109 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.788957 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-config-data" (OuterVolumeSpecName: "config-data") pod "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" (UID: "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.809612 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" (UID: "c3c570a4-a49e-4c7f-9bc4-54d7569c87e7"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.889690 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-combined-ca-bundle\") pod \"f1a51fac-9422-488c-922a-c42e21ccd7df\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.889803 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kch9r\" (UniqueName: \"kubernetes.io/projected/f1a51fac-9422-488c-922a-c42e21ccd7df-kube-api-access-kch9r\") pod \"f1a51fac-9422-488c-922a-c42e21ccd7df\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.889869 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a51fac-9422-488c-922a-c42e21ccd7df-logs\") pod \"f1a51fac-9422-488c-922a-c42e21ccd7df\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.889928 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-config-data\") pod \"f1a51fac-9422-488c-922a-c42e21ccd7df\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.889977 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-cert-memcached-mtls\") pod \"f1a51fac-9422-488c-922a-c42e21ccd7df\" (UID: \"f1a51fac-9422-488c-922a-c42e21ccd7df\") " Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.890295 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.890313 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.890329 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1a51fac-9422-488c-922a-c42e21ccd7df-logs" (OuterVolumeSpecName: "logs") pod "f1a51fac-9422-488c-922a-c42e21ccd7df" (UID: "f1a51fac-9422-488c-922a-c42e21ccd7df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.893242 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1a51fac-9422-488c-922a-c42e21ccd7df-kube-api-access-kch9r" (OuterVolumeSpecName: "kube-api-access-kch9r") pod "f1a51fac-9422-488c-922a-c42e21ccd7df" (UID: "f1a51fac-9422-488c-922a-c42e21ccd7df"). InnerVolumeSpecName "kube-api-access-kch9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.917620 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1a51fac-9422-488c-922a-c42e21ccd7df" (UID: "f1a51fac-9422-488c-922a-c42e21ccd7df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.940351 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-config-data" (OuterVolumeSpecName: "config-data") pod "f1a51fac-9422-488c-922a-c42e21ccd7df" (UID: "f1a51fac-9422-488c-922a-c42e21ccd7df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.950168 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "f1a51fac-9422-488c-922a-c42e21ccd7df" (UID: "f1a51fac-9422-488c-922a-c42e21ccd7df"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.991780 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.991828 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kch9r\" (UniqueName: \"kubernetes.io/projected/f1a51fac-9422-488c-922a-c42e21ccd7df-kube-api-access-kch9r\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.991846 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a51fac-9422-488c-922a-c42e21ccd7df-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.991859 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:12 crc kubenswrapper[4807]: I1205 12:38:12.991871 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f1a51fac-9422-488c-922a-c42e21ccd7df-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.091237 4807 generic.go:334] "Generic (PLEG): container finished" podID="2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9" containerID="fedbebd8ae752c9fa6d9d23f8d3a5b5c3fb6eccda16303307b316631ec932efe" exitCode=0 Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.091318 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" event={"ID":"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9","Type":"ContainerDied","Data":"fedbebd8ae752c9fa6d9d23f8d3a5b5c3fb6eccda16303307b316631ec932efe"} Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.092994 4807 generic.go:334] "Generic (PLEG): container finished" podID="f1a51fac-9422-488c-922a-c42e21ccd7df" containerID="4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e" exitCode=0 Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.093041 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.093056 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"f1a51fac-9422-488c-922a-c42e21ccd7df","Type":"ContainerDied","Data":"4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e"} Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.093078 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"f1a51fac-9422-488c-922a-c42e21ccd7df","Type":"ContainerDied","Data":"45ae363cd449d04b185adaef691d161424829e7d1ad2f3506ea5cdcd3f652e23"} Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.093096 4807 scope.go:117] "RemoveContainer" containerID="4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.097134 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"c3c570a4-a49e-4c7f-9bc4-54d7569c87e7","Type":"ContainerDied","Data":"2d4337a6b57e2021986af3b857969d76dc8a42c6797f3452d00a49dfeaf5830d"} Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.097223 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.139198 4807 scope.go:117] "RemoveContainer" containerID="4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.141022 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:38:13 crc kubenswrapper[4807]: E1205 12:38:13.141314 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e\": container with ID starting with 4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e not found: ID does not exist" containerID="4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.141352 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e"} err="failed to get container status \"4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e\": rpc error: code = NotFound desc = could not find container \"4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e\": container with ID starting with 4b622c876a29f334d9ba8b025d116289ac4d580647043249ad8c68608012ce5e not found: ID does not exist" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.141382 4807 scope.go:117] "RemoveContainer" containerID="32958eadddc2a80e25cbd731c2470867f6b5e87c8751430309994021eb0bfa48" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.157602 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.164068 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.166706 4807 scope.go:117] "RemoveContainer" containerID="2338f381c9e544e5774aaa38220b2ef7ce00304710f06439b086c7159712abaf" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.170105 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.254707 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" path="/var/lib/kubelet/pods/c3c570a4-a49e-4c7f-9bc4-54d7569c87e7/volumes" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.255465 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" path="/var/lib/kubelet/pods/e849db0d-4bb4-48aa-bb8f-8730d424694b/volumes" Dec 05 12:38:13 crc kubenswrapper[4807]: I1205 12:38:13.256132 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1a51fac-9422-488c-922a-c42e21ccd7df" path="/var/lib/kubelet/pods/f1a51fac-9422-488c-922a-c42e21ccd7df/volumes" Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.462301 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.616919 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np8gq\" (UniqueName: \"kubernetes.io/projected/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-kube-api-access-np8gq\") pod \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.617173 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-operator-scripts\") pod \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\" (UID: \"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9\") " Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.617881 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9" (UID: "2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.629296 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-kube-api-access-np8gq" (OuterVolumeSpecName: "kube-api-access-np8gq") pod "2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9" (UID: "2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9"). InnerVolumeSpecName "kube-api-access-np8gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.718999 4807 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.719032 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np8gq\" (UniqueName: \"kubernetes.io/projected/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9-kube-api-access-np8gq\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.800818 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.801244 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-notification-agent" containerID="cri-o://d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db" gracePeriod=30 Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.801266 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="proxy-httpd" containerID="cri-o://ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965" gracePeriod=30 Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.801270 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="sg-core" containerID="cri-o://a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e" gracePeriod=30 Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.801167 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-central-agent" containerID="cri-o://7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5" gracePeriod=30 Dec 05 12:38:14 crc kubenswrapper[4807]: I1205 12:38:14.807750 4807 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.224:3000/\": EOF" Dec 05 12:38:15 crc kubenswrapper[4807]: I1205 12:38:15.139194 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" event={"ID":"2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9","Type":"ContainerDied","Data":"e8f4077c43553a2d793097a71fd264968c4904bb69df59dc0e0694d93e606976"} Dec 05 12:38:15 crc kubenswrapper[4807]: I1205 12:38:15.139248 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8f4077c43553a2d793097a71fd264968c4904bb69df59dc0e0694d93e606976" Dec 05 12:38:15 crc kubenswrapper[4807]: I1205 12:38:15.139321 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-sr7cp" Dec 05 12:38:15 crc kubenswrapper[4807]: I1205 12:38:15.145569 4807 generic.go:334] "Generic (PLEG): container finished" podID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerID="ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965" exitCode=0 Dec 05 12:38:15 crc kubenswrapper[4807]: I1205 12:38:15.145615 4807 generic.go:334] "Generic (PLEG): container finished" podID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerID="a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e" exitCode=2 Dec 05 12:38:15 crc kubenswrapper[4807]: I1205 12:38:15.145640 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerDied","Data":"ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965"} Dec 05 12:38:15 crc kubenswrapper[4807]: I1205 12:38:15.145669 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerDied","Data":"a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e"} Dec 05 12:38:15 crc kubenswrapper[4807]: E1205 12:38:15.265509 4807 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dbc383e_bfef_4869_b8cf_e71a4cfa5ba9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58286ab5_dd9b_463c_95af_5ae2872ed8b4.slice/crio-conmon-ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dbc383e_bfef_4869_b8cf_e71a4cfa5ba9.slice/crio-e8f4077c43553a2d793097a71fd264968c4904bb69df59dc0e0694d93e606976\": RecentStats: unable to find data in memory cache]" Dec 05 12:38:16 crc kubenswrapper[4807]: I1205 12:38:16.155739 4807 generic.go:334] "Generic (PLEG): container finished" podID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerID="7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5" exitCode=0 Dec 05 12:38:16 crc kubenswrapper[4807]: I1205 12:38:16.155818 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerDied","Data":"7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5"} Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.695619 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808183 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-log-httpd\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808279 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-config-data\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808317 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-combined-ca-bundle\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808361 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-run-httpd\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808412 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crzjw\" (UniqueName: \"kubernetes.io/projected/58286ab5-dd9b-463c-95af-5ae2872ed8b4-kube-api-access-crzjw\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808461 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-sg-core-conf-yaml\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808487 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-ceilometer-tls-certs\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808533 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-scripts\") pod \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\" (UID: \"58286ab5-dd9b-463c-95af-5ae2872ed8b4\") " Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.808780 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.809057 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.817784 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-scripts" (OuterVolumeSpecName: "scripts") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.833955 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58286ab5-dd9b-463c-95af-5ae2872ed8b4-kube-api-access-crzjw" (OuterVolumeSpecName: "kube-api-access-crzjw") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "kube-api-access-crzjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.838706 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.890190 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.917341 4807 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.917380 4807 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.917395 4807 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.917405 4807 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.917415 4807 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58286ab5-dd9b-463c-95af-5ae2872ed8b4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.917425 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crzjw\" (UniqueName: \"kubernetes.io/projected/58286ab5-dd9b-463c-95af-5ae2872ed8b4-kube-api-access-crzjw\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.931860 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:19 crc kubenswrapper[4807]: I1205 12:38:19.954792 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-config-data" (OuterVolumeSpecName: "config-data") pod "58286ab5-dd9b-463c-95af-5ae2872ed8b4" (UID: "58286ab5-dd9b-463c-95af-5ae2872ed8b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.019292 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.019684 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58286ab5-dd9b-463c-95af-5ae2872ed8b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.022706 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.032026 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-7gt2n"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.041946 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-7gt2n"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.071980 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-sr7cp"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.079322 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-tbdf5"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.087218 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-sr7cp"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.094108 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-tbdf5"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.120481 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-combined-ca-bundle\") pod \"88b4b699-36f9-4259-9fb6-f241f12550c9\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.120812 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-config-data\") pod \"88b4b699-36f9-4259-9fb6-f241f12550c9\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.120848 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88b4b699-36f9-4259-9fb6-f241f12550c9-logs\") pod \"88b4b699-36f9-4259-9fb6-f241f12550c9\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.120944 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-cert-memcached-mtls\") pod \"88b4b699-36f9-4259-9fb6-f241f12550c9\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.120984 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-custom-prometheus-ca\") pod \"88b4b699-36f9-4259-9fb6-f241f12550c9\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.121015 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5n5b\" (UniqueName: \"kubernetes.io/projected/88b4b699-36f9-4259-9fb6-f241f12550c9-kube-api-access-g5n5b\") pod \"88b4b699-36f9-4259-9fb6-f241f12550c9\" (UID: \"88b4b699-36f9-4259-9fb6-f241f12550c9\") " Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.121319 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88b4b699-36f9-4259-9fb6-f241f12550c9-logs" (OuterVolumeSpecName: "logs") pod "88b4b699-36f9-4259-9fb6-f241f12550c9" (UID: "88b4b699-36f9-4259-9fb6-f241f12550c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.121447 4807 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88b4b699-36f9-4259-9fb6-f241f12550c9-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.124194 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b4b699-36f9-4259-9fb6-f241f12550c9-kube-api-access-g5n5b" (OuterVolumeSpecName: "kube-api-access-g5n5b") pod "88b4b699-36f9-4259-9fb6-f241f12550c9" (UID: "88b4b699-36f9-4259-9fb6-f241f12550c9"). InnerVolumeSpecName "kube-api-access-g5n5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.140921 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88b4b699-36f9-4259-9fb6-f241f12550c9" (UID: "88b4b699-36f9-4259-9fb6-f241f12550c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.141802 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "88b4b699-36f9-4259-9fb6-f241f12550c9" (UID: "88b4b699-36f9-4259-9fb6-f241f12550c9"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.161736 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-config-data" (OuterVolumeSpecName: "config-data") pod "88b4b699-36f9-4259-9fb6-f241f12550c9" (UID: "88b4b699-36f9-4259-9fb6-f241f12550c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.179093 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "88b4b699-36f9-4259-9fb6-f241f12550c9" (UID: "88b4b699-36f9-4259-9fb6-f241f12550c9"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.193758 4807 generic.go:334] "Generic (PLEG): container finished" podID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerID="d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db" exitCode=0 Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.193824 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.193846 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerDied","Data":"d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db"} Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.193955 4807 scope.go:117] "RemoveContainer" containerID="ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.194215 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"58286ab5-dd9b-463c-95af-5ae2872ed8b4","Type":"ContainerDied","Data":"ee8c3dbdb3cc60a8324d721a50515a63c4c18787411acda6d4277323755e0522"} Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.196233 4807 generic.go:334] "Generic (PLEG): container finished" podID="88b4b699-36f9-4259-9fb6-f241f12550c9" containerID="2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9" exitCode=0 Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.196254 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"88b4b699-36f9-4259-9fb6-f241f12550c9","Type":"ContainerDied","Data":"2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9"} Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.196268 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"88b4b699-36f9-4259-9fb6-f241f12550c9","Type":"ContainerDied","Data":"5dc7ede7ced29bc29a5d9c917f2857211d9e8ea15deaede18c619646eb45e563"} Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.196315 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.215686 4807 scope.go:117] "RemoveContainer" containerID="a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.222681 4807 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.222712 4807 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.222721 4807 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.222732 4807 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/88b4b699-36f9-4259-9fb6-f241f12550c9-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.222741 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5n5b\" (UniqueName: \"kubernetes.io/projected/88b4b699-36f9-4259-9fb6-f241f12550c9-kube-api-access-g5n5b\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.300319 4807 scope.go:117] "RemoveContainer" containerID="d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.308890 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.317281 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.325969 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.334234 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.339822 4807 scope.go:117] "RemoveContainer" containerID="7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.343288 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.343845 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-api" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.344076 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-api" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.344413 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-api" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.344501 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-api" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.344830 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="sg-core" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.344903 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="sg-core" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.344974 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9" containerName="mariadb-account-delete" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.345058 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9" containerName="mariadb-account-delete" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.345157 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b4b699-36f9-4259-9fb6-f241f12550c9" containerName="watcher-decision-engine" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.345231 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b4b699-36f9-4259-9fb6-f241f12550c9" containerName="watcher-decision-engine" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.345314 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-notification-agent" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.345399 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-notification-agent" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.345477 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="proxy-httpd" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.345567 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="proxy-httpd" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.345635 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-kuttl-api-log" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.345687 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-kuttl-api-log" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.345752 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a51fac-9422-488c-922a-c42e21ccd7df" containerName="watcher-applier" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.345817 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a51fac-9422-488c-922a-c42e21ccd7df" containerName="watcher-applier" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.345892 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-kuttl-api-log" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.345965 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-kuttl-api-log" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.346093 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-central-agent" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.346174 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-central-agent" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.346453 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-notification-agent" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.347793 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-api" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.347945 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-api" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348037 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b4b699-36f9-4259-9fb6-f241f12550c9" containerName="watcher-decision-engine" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348132 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1a51fac-9422-488c-922a-c42e21ccd7df" containerName="watcher-applier" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348209 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="proxy-httpd" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348282 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="sg-core" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348346 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c570a4-a49e-4c7f-9bc4-54d7569c87e7" containerName="watcher-kuttl-api-log" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348407 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9" containerName="mariadb-account-delete" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348465 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="e849db0d-4bb4-48aa-bb8f-8730d424694b" containerName="watcher-kuttl-api-log" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.348555 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" containerName="ceilometer-central-agent" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.350262 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.350472 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.357983 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.358288 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.358444 4807 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.359880 4807 scope.go:117] "RemoveContainer" containerID="ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.363770 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965\": container with ID starting with ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965 not found: ID does not exist" containerID="ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.363809 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965"} err="failed to get container status \"ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965\": rpc error: code = NotFound desc = could not find container \"ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965\": container with ID starting with ab41603a0a85136fc851f1b4fdd2d515650512cbcfd8368e03c68ddddaad0965 not found: ID does not exist" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.363836 4807 scope.go:117] "RemoveContainer" containerID="a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.364074 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e\": container with ID starting with a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e not found: ID does not exist" containerID="a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.364094 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e"} err="failed to get container status \"a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e\": rpc error: code = NotFound desc = could not find container \"a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e\": container with ID starting with a43a9e757b6ea84932dab187cb459e7004e768e94ad03d885d145c9d707a499e not found: ID does not exist" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.364113 4807 scope.go:117] "RemoveContainer" containerID="d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.364360 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db\": container with ID starting with d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db not found: ID does not exist" containerID="d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.364384 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db"} err="failed to get container status \"d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db\": rpc error: code = NotFound desc = could not find container \"d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db\": container with ID starting with d92078ab62bae63585474cd138537c9a91c88db57fb76fa1299be673c3e2c5db not found: ID does not exist" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.364404 4807 scope.go:117] "RemoveContainer" containerID="7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.364782 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5\": container with ID starting with 7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5 not found: ID does not exist" containerID="7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.364811 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5"} err="failed to get container status \"7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5\": rpc error: code = NotFound desc = could not find container \"7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5\": container with ID starting with 7218bf3921a39848cc133557f100fede4c9f709e199832e1237a88fdac57f3a5 not found: ID does not exist" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.364828 4807 scope.go:117] "RemoveContainer" containerID="2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.384456 4807 scope.go:117] "RemoveContainer" containerID="2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9" Dec 05 12:38:20 crc kubenswrapper[4807]: E1205 12:38:20.386938 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9\": container with ID starting with 2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9 not found: ID does not exist" containerID="2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.386982 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9"} err="failed to get container status \"2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9\": rpc error: code = NotFound desc = could not find container \"2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9\": container with ID starting with 2419a61618bc48f022adcc8e339f90baddf22db85483fe1ccb3126e7451399f9 not found: ID does not exist" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528078 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528135 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-config-data\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528175 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db846f0a-96f1-4a17-bdcc-b65bb02094b2-run-httpd\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528199 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-scripts\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528219 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8cs8\" (UniqueName: \"kubernetes.io/projected/db846f0a-96f1-4a17-bdcc-b65bb02094b2-kube-api-access-l8cs8\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528354 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528437 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db846f0a-96f1-4a17-bdcc-b65bb02094b2-log-httpd\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.528457 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629545 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-config-data\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629585 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db846f0a-96f1-4a17-bdcc-b65bb02094b2-run-httpd\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629605 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-scripts\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629621 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8cs8\" (UniqueName: \"kubernetes.io/projected/db846f0a-96f1-4a17-bdcc-b65bb02094b2-kube-api-access-l8cs8\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629665 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629715 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db846f0a-96f1-4a17-bdcc-b65bb02094b2-log-httpd\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629733 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.629770 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.630230 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db846f0a-96f1-4a17-bdcc-b65bb02094b2-run-httpd\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.630510 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db846f0a-96f1-4a17-bdcc-b65bb02094b2-log-httpd\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.633934 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-scripts\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.633988 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.634169 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-config-data\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.635011 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.635564 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db846f0a-96f1-4a17-bdcc-b65bb02094b2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.648196 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8cs8\" (UniqueName: \"kubernetes.io/projected/db846f0a-96f1-4a17-bdcc-b65bb02094b2-kube-api-access-l8cs8\") pod \"ceilometer-0\" (UID: \"db846f0a-96f1-4a17-bdcc-b65bb02094b2\") " pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:20 crc kubenswrapper[4807]: I1205 12:38:20.670307 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:21 crc kubenswrapper[4807]: I1205 12:38:21.125795 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Dec 05 12:38:21 crc kubenswrapper[4807]: I1205 12:38:21.232664 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"db846f0a-96f1-4a17-bdcc-b65bb02094b2","Type":"ContainerStarted","Data":"ae1277577c9467dd9e01248b82efb37e43bc69ea373eda89a3803d634c920ee7"} Dec 05 12:38:21 crc kubenswrapper[4807]: I1205 12:38:21.262092 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9" path="/var/lib/kubelet/pods/2dbc383e-bfef-4869-b8cf-e71a4cfa5ba9/volumes" Dec 05 12:38:21 crc kubenswrapper[4807]: I1205 12:38:21.263071 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58286ab5-dd9b-463c-95af-5ae2872ed8b4" path="/var/lib/kubelet/pods/58286ab5-dd9b-463c-95af-5ae2872ed8b4/volumes" Dec 05 12:38:21 crc kubenswrapper[4807]: I1205 12:38:21.263870 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68413519-afb4-4f45-97e0-54b989b13fd6" path="/var/lib/kubelet/pods/68413519-afb4-4f45-97e0-54b989b13fd6/volumes" Dec 05 12:38:21 crc kubenswrapper[4807]: I1205 12:38:21.265468 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b4b699-36f9-4259-9fb6-f241f12550c9" path="/var/lib/kubelet/pods/88b4b699-36f9-4259-9fb6-f241f12550c9/volumes" Dec 05 12:38:21 crc kubenswrapper[4807]: I1205 12:38:21.266192 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dac84884-07d6-4bc5-aa0d-a7f1dd579a93" path="/var/lib/kubelet/pods/dac84884-07d6-4bc5-aa0d-a7f1dd579a93/volumes" Dec 05 12:38:23 crc kubenswrapper[4807]: I1205 12:38:23.283827 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"db846f0a-96f1-4a17-bdcc-b65bb02094b2","Type":"ContainerStarted","Data":"133d2e0c5d2916ff1dc299b102086ee3c627ca47dcba696e8e74069ad77d388b"} Dec 05 12:38:24 crc kubenswrapper[4807]: I1205 12:38:24.298915 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"db846f0a-96f1-4a17-bdcc-b65bb02094b2","Type":"ContainerStarted","Data":"db1843de919b83d2a6f87722fdf58b7a9386bf5efbcbafd7ada98bfc95f6ae9d"} Dec 05 12:38:24 crc kubenswrapper[4807]: I1205 12:38:24.299254 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"db846f0a-96f1-4a17-bdcc-b65bb02094b2","Type":"ContainerStarted","Data":"3f6582dc981a97d7315198cb722ce26ea759f23fe635eca5f76ac541acbdd459"} Dec 05 12:38:26 crc kubenswrapper[4807]: I1205 12:38:26.316024 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"db846f0a-96f1-4a17-bdcc-b65bb02094b2","Type":"ContainerStarted","Data":"a9a16dc5540fe63753455d5ada90dc2c72fe206305acbc83c78ca2d2b65701b5"} Dec 05 12:38:26 crc kubenswrapper[4807]: I1205 12:38:26.316586 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:26 crc kubenswrapper[4807]: I1205 12:38:26.339838 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.882569537 podStartE2EDuration="6.339801475s" podCreationTimestamp="2025-12-05 12:38:20 +0000 UTC" firstStartedPulling="2025-12-05 12:38:21.133112369 +0000 UTC m=+1930.626975638" lastFinishedPulling="2025-12-05 12:38:25.590344317 +0000 UTC m=+1935.084207576" observedRunningTime="2025-12-05 12:38:26.338990306 +0000 UTC m=+1935.832853585" watchObservedRunningTime="2025-12-05 12:38:26.339801475 +0000 UTC m=+1935.833664744" Dec 05 12:38:29 crc kubenswrapper[4807]: I1205 12:38:29.906065 4807 scope.go:117] "RemoveContainer" containerID="56d4f3dc25a1baabfa69697662d7947396c5b37052889719cb8740cd7db4b654" Dec 05 12:38:29 crc kubenswrapper[4807]: I1205 12:38:29.961784 4807 scope.go:117] "RemoveContainer" containerID="d310361021ce548f3a27d4d0ad3409df1ad5a47e895739cd5e779a37004cee08" Dec 05 12:38:30 crc kubenswrapper[4807]: I1205 12:38:30.013909 4807 scope.go:117] "RemoveContainer" containerID="b3a5525bfa1b8468b7b4468b15fa875ad0647bb165b67b8e04b37df5cec635cf" Dec 05 12:38:30 crc kubenswrapper[4807]: I1205 12:38:30.054561 4807 scope.go:117] "RemoveContainer" containerID="cb9d46c5057b570f2d7178d833eac2de5e7af8162e538e28de8c43d818fb4866" Dec 05 12:38:30 crc kubenswrapper[4807]: I1205 12:38:30.076494 4807 scope.go:117] "RemoveContainer" containerID="eb2aafa140e8b2537becc06f69ead552cf44fe114bf70966e0de9a32ffa41f73" Dec 05 12:38:50 crc kubenswrapper[4807]: I1205 12:38:50.677837 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.442230 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zm9tk/must-gather-8lb9h"] Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.444948 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.448442 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zm9tk"/"openshift-service-ca.crt" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.451071 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zm9tk"/"kube-root-ca.crt" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.459635 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zm9tk"/"default-dockercfg-p2trd" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.469704 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zm9tk/must-gather-8lb9h"] Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.644080 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4cfx\" (UniqueName: \"kubernetes.io/projected/a1da4dc7-4134-414d-8494-bf7bd296d80f-kube-api-access-t4cfx\") pod \"must-gather-8lb9h\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.644148 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a1da4dc7-4134-414d-8494-bf7bd296d80f-must-gather-output\") pod \"must-gather-8lb9h\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.745784 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a1da4dc7-4134-414d-8494-bf7bd296d80f-must-gather-output\") pod \"must-gather-8lb9h\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.745930 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4cfx\" (UniqueName: \"kubernetes.io/projected/a1da4dc7-4134-414d-8494-bf7bd296d80f-kube-api-access-t4cfx\") pod \"must-gather-8lb9h\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.746312 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a1da4dc7-4134-414d-8494-bf7bd296d80f-must-gather-output\") pod \"must-gather-8lb9h\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.765488 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4cfx\" (UniqueName: \"kubernetes.io/projected/a1da4dc7-4134-414d-8494-bf7bd296d80f-kube-api-access-t4cfx\") pod \"must-gather-8lb9h\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:58 crc kubenswrapper[4807]: I1205 12:38:58.768638 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:38:59 crc kubenswrapper[4807]: I1205 12:38:59.254350 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zm9tk/must-gather-8lb9h"] Dec 05 12:38:59 crc kubenswrapper[4807]: I1205 12:38:59.268881 4807 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:38:59 crc kubenswrapper[4807]: I1205 12:38:59.580097 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" event={"ID":"a1da4dc7-4134-414d-8494-bf7bd296d80f","Type":"ContainerStarted","Data":"85b83993595f81f0de23bcc60ba6dc316f9678470618ee02334831de762a0e31"} Dec 05 12:39:04 crc kubenswrapper[4807]: I1205 12:39:04.622943 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" event={"ID":"a1da4dc7-4134-414d-8494-bf7bd296d80f","Type":"ContainerStarted","Data":"902a9252892d45cd26292d61bc1710505d9a5fc057b91241a6206ac39d089bc8"} Dec 05 12:39:04 crc kubenswrapper[4807]: I1205 12:39:04.624008 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" event={"ID":"a1da4dc7-4134-414d-8494-bf7bd296d80f","Type":"ContainerStarted","Data":"0c46ae024a3db1eb2624c1331f1927076c0edf3a4bcbbeaf418722b3b18874fc"} Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.031739 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl_5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6/util/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.215161 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl_5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6/pull/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.236300 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl_5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6/util/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.271838 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl_5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6/pull/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.418316 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl_5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6/pull/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.446414 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl_5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6/extract/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.448178 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2aca67bceaf5f8f951e258252dcf1c8090dc81397508fd629a1381c43fcg7jl_5828ffd4-1cdb-4474-88cf-a6b1ad1a78f6/util/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.595062 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq_652de9e9-c06e-4661-81f6-8dec66fa510f/util/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.814336 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq_652de9e9-c06e-4661-81f6-8dec66fa510f/pull/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.824051 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq_652de9e9-c06e-4661-81f6-8dec66fa510f/pull/0.log" Dec 05 12:40:09 crc kubenswrapper[4807]: I1205 12:40:09.850800 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq_652de9e9-c06e-4661-81f6-8dec66fa510f/util/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.062512 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq_652de9e9-c06e-4661-81f6-8dec66fa510f/pull/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.093724 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq_652de9e9-c06e-4661-81f6-8dec66fa510f/util/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.098387 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b3d56a81a1998265346272a04a98d31ccec6e2ed9a681290e577d2c515kzrq_652de9e9-c06e-4661-81f6-8dec66fa510f/extract/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.355444 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-zjzsd_d2b94a74-f984-4360-8062-e42b82a7c401/kube-rbac-proxy/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.381805 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-zjzsd_d2b94a74-f984-4360-8062-e42b82a7c401/manager/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.404156 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ng4kv_022ffaeb-bad6-48ee-be09-78fc4c515a99/kube-rbac-proxy/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.579030 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ng4kv_022ffaeb-bad6-48ee-be09-78fc4c515a99/manager/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.639537 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-t82v9_d36f8a3a-450b-4824-826c-32283863f270/kube-rbac-proxy/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.648457 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-t82v9_d36f8a3a-450b-4824-826c-32283863f270/manager/0.log" Dec 05 12:40:10 crc kubenswrapper[4807]: I1205 12:40:10.848406 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-knvh4_c383d37a-acb0-463c-9980-1d2d46e06291/kube-rbac-proxy/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.072269 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-knvh4_c383d37a-acb0-463c-9980-1d2d46e06291/manager/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.224672 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-j26nl_ef9d3954-d526-4bd6-b806-e0ddb31994d3/manager/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.312872 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-blrcs_7cd89edb-31e7-4c50-ba4d-c86f85fce855/kube-rbac-proxy/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.317772 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-j26nl_ef9d3954-d526-4bd6-b806-e0ddb31994d3/kube-rbac-proxy/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.422770 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-blrcs_7cd89edb-31e7-4c50-ba4d-c86f85fce855/manager/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.517171 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-v6m8x_4a8a470d-bdcf-4d64-b805-c996e88384ae/kube-rbac-proxy/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.742460 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-v6m8x_4a8a470d-bdcf-4d64-b805-c996e88384ae/manager/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.757893 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7pszd_5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86/kube-rbac-proxy/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.855582 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7pszd_5cb231c4-a9c8-4e5c-a483-45d5b3aa3f86/manager/0.log" Dec 05 12:40:11 crc kubenswrapper[4807]: I1205 12:40:11.944188 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-vl87h_64226f84-d145-459b-b326-acf7e87add59/kube-rbac-proxy/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.049056 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-vl87h_64226f84-d145-459b-b326-acf7e87add59/manager/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.165688 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-5qfr7_65950bcd-6c98-49b5-b25f-f58cd0ee72bd/kube-rbac-proxy/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.214775 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-5qfr7_65950bcd-6c98-49b5-b25f-f58cd0ee72bd/manager/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.297634 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-g5tvq_8817b3f7-5a96-42e2-9807-38a6742f5ac3/kube-rbac-proxy/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.367513 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-g5tvq_8817b3f7-5a96-42e2-9807-38a6742f5ac3/manager/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.498838 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rzhvm_568f4286-ea33-4e00-ab87-cdc930373223/manager/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.514043 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rzhvm_568f4286-ea33-4e00-ab87-cdc930373223/kube-rbac-proxy/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.698093 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jzdm6_64763a6c-60c1-428e-baf9-040111b0057e/kube-rbac-proxy/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.773574 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jzdm6_64763a6c-60c1-428e-baf9-040111b0057e/manager/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.803348 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-t8bz9_49f023c0-dfc7-4035-ae05-df3e6913e581/kube-rbac-proxy/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.922698 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-t8bz9_49f023c0-dfc7-4035-ae05-df3e6913e581/manager/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.979412 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8_f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a/kube-rbac-proxy/0.log" Dec 05 12:40:12 crc kubenswrapper[4807]: I1205 12:40:12.986372 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4ll6z8_f6a39a74-f2be-48ff-8cb3-56c7a0dd0b4a/manager/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.211331 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-x89f7_89e92216-a04d-416c-84b1-fbaa6e2e52b5/registry-server/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.309566 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-99xgh_336d4305-fc09-4622-b6fd-ebf4fe266ad2/kube-rbac-proxy/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.458719 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-99xgh_336d4305-fc09-4622-b6fd-ebf4fe266ad2/manager/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.474436 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-fcb8bd8db-sbftd_bc48b693-8a75-4dea-a3cf-3dd053a8a500/manager/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.502806 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nclwx_2f6ae961-a4b7-4d96-8669-2b0c1653ddc3/kube-rbac-proxy/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.567904 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nclwx_2f6ae961-a4b7-4d96-8669-2b0c1653ddc3/manager/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.711463 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-r2q7v_87b435a4-0ca7-4a37-9736-40ff511bfb43/operator/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.752403 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-zbdbh_85952cd3-56c3-49d2-9f15-cd1f872a6c7c/kube-rbac-proxy/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.800173 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-zbdbh_85952cd3-56c3-49d2-9f15-cd1f872a6c7c/manager/0.log" Dec 05 12:40:13 crc kubenswrapper[4807]: I1205 12:40:13.968281 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-mfzmp_63696f3a-63f2-475e-b5e1-9ed44bc6eb39/kube-rbac-proxy/0.log" Dec 05 12:40:14 crc kubenswrapper[4807]: I1205 12:40:14.047869 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-g2mx8_5be4b980-a6ab-4499-87ec-0aff47a7a917/kube-rbac-proxy/0.log" Dec 05 12:40:14 crc kubenswrapper[4807]: I1205 12:40:14.171324 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-mfzmp_63696f3a-63f2-475e-b5e1-9ed44bc6eb39/manager/0.log" Dec 05 12:40:14 crc kubenswrapper[4807]: I1205 12:40:14.208759 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-g2mx8_5be4b980-a6ab-4499-87ec-0aff47a7a917/manager/0.log" Dec 05 12:40:14 crc kubenswrapper[4807]: I1205 12:40:14.518615 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-index-vx7jt_49dfc8b3-c6d0-4c3f-a056-f6e4180068bd/registry-server/0.log" Dec 05 12:40:14 crc kubenswrapper[4807]: I1205 12:40:14.603352 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7b5d5fdf58-88nkt_ae6e3e1b-30de-427e-9616-8d023b6b97c5/manager/0.log" Dec 05 12:40:22 crc kubenswrapper[4807]: I1205 12:40:22.466543 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:40:22 crc kubenswrapper[4807]: I1205 12:40:22.467090 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:40:30 crc kubenswrapper[4807]: I1205 12:40:30.357023 4807 scope.go:117] "RemoveContainer" containerID="178e105ec129899fd78b2dc51dcdbbed1335d1980a83ec82b2fd0c37caf08286" Dec 05 12:40:30 crc kubenswrapper[4807]: I1205 12:40:30.391752 4807 scope.go:117] "RemoveContainer" containerID="60c7c8c7cd2f87db9f4e08ad775231377716b19983a5a65957bcf3909cff53f7" Dec 05 12:40:30 crc kubenswrapper[4807]: I1205 12:40:30.422734 4807 scope.go:117] "RemoveContainer" containerID="d8f0f5ffc9b8753c94122ddf3f3ac4650009d358ffe8ead4664c118401ee1833" Dec 05 12:40:30 crc kubenswrapper[4807]: I1205 12:40:30.458680 4807 scope.go:117] "RemoveContainer" containerID="79e5a776728cf67d2ffa0d5beb3b612daea3485a2ca450ae59b037e174fec324" Dec 05 12:40:36 crc kubenswrapper[4807]: I1205 12:40:36.213929 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r9cjd_a0ddd806-8a2b-4cc5-9399-cb3f86eab1e6/control-plane-machine-set-operator/0.log" Dec 05 12:40:36 crc kubenswrapper[4807]: I1205 12:40:36.435162 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xrqgd_b03b24c9-59d5-45cd-8416-7ee0d0b3f08e/machine-api-operator/0.log" Dec 05 12:40:36 crc kubenswrapper[4807]: I1205 12:40:36.529160 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xrqgd_b03b24c9-59d5-45cd-8416-7ee0d0b3f08e/kube-rbac-proxy/0.log" Dec 05 12:40:50 crc kubenswrapper[4807]: I1205 12:40:50.362725 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-gh6vh_754c38f5-b65c-4ec6-905e-b5e1a3ca6cec/cert-manager-controller/0.log" Dec 05 12:40:50 crc kubenswrapper[4807]: I1205 12:40:50.607224 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-65hxg_82008b0e-6099-474a-8526-f531424944ff/cert-manager-webhook/0.log" Dec 05 12:40:50 crc kubenswrapper[4807]: I1205 12:40:50.656415 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-tvt8k_574e8b05-7725-48f4-afdf-381468494129/cert-manager-cainjector/0.log" Dec 05 12:40:52 crc kubenswrapper[4807]: I1205 12:40:52.466556 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:40:52 crc kubenswrapper[4807]: I1205 12:40:52.466618 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:41:04 crc kubenswrapper[4807]: I1205 12:41:04.564223 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-fm2ks_79eac56e-1d44-4f6b-875a-8f944ee63f7a/nmstate-console-plugin/0.log" Dec 05 12:41:04 crc kubenswrapper[4807]: I1205 12:41:04.806438 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-hp4qg_1ef69c53-a12b-4ae8-ae6f-baa299acfb6a/nmstate-handler/0.log" Dec 05 12:41:04 crc kubenswrapper[4807]: I1205 12:41:04.843104 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p442k_222fa6f9-b50a-4ef2-82fd-7b9d37841def/kube-rbac-proxy/0.log" Dec 05 12:41:04 crc kubenswrapper[4807]: I1205 12:41:04.903156 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p442k_222fa6f9-b50a-4ef2-82fd-7b9d37841def/nmstate-metrics/0.log" Dec 05 12:41:05 crc kubenswrapper[4807]: I1205 12:41:05.107311 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-kbls9_647cdbda-7377-42a5-80d2-d67f6babcc58/nmstate-operator/0.log" Dec 05 12:41:05 crc kubenswrapper[4807]: I1205 12:41:05.213552 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-tcgqr_6d436024-c16c-4428-81db-a144ed2deecc/nmstate-webhook/0.log" Dec 05 12:41:21 crc kubenswrapper[4807]: I1205 12:41:21.465170 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-52sxs_26eb0077-d64d-4b66-90ee-9e994c94916f/kube-rbac-proxy/0.log" Dec 05 12:41:21 crc kubenswrapper[4807]: I1205 12:41:21.620992 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-52sxs_26eb0077-d64d-4b66-90ee-9e994c94916f/controller/0.log" Dec 05 12:41:21 crc kubenswrapper[4807]: I1205 12:41:21.762816 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-frr-files/0.log" Dec 05 12:41:21 crc kubenswrapper[4807]: I1205 12:41:21.955279 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-reloader/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.022625 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-metrics/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.025167 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-frr-files/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.052775 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-reloader/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.280632 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-frr-files/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.283303 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-metrics/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.328518 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-metrics/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.388459 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-reloader/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.466226 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.466284 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.466331 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.466954 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d1db79e77bd059c1448ecb5528526ada5577ec3cff58469fb098ce90654d049"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.467012 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://5d1db79e77bd059c1448ecb5528526ada5577ec3cff58469fb098ce90654d049" gracePeriod=600 Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.571879 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-reloader/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.583817 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-metrics/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.599978 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/cp-frr-files/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.606887 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/controller/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.744623 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="5d1db79e77bd059c1448ecb5528526ada5577ec3cff58469fb098ce90654d049" exitCode=0 Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.744686 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"5d1db79e77bd059c1448ecb5528526ada5577ec3cff58469fb098ce90654d049"} Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.744727 4807 scope.go:117] "RemoveContainer" containerID="3e433667f105a493fbbd1996e9c48980a351faf557c4bc9f9f92376698d3e79e" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.786659 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/kube-rbac-proxy/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.846535 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/kube-rbac-proxy-frr/0.log" Dec 05 12:41:22 crc kubenswrapper[4807]: I1205 12:41:22.915920 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/frr-metrics/0.log" Dec 05 12:41:23 crc kubenswrapper[4807]: I1205 12:41:23.064779 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/reloader/0.log" Dec 05 12:41:23 crc kubenswrapper[4807]: I1205 12:41:23.433386 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-6wtk4_b144d4af-48d9-455d-8e32-3f5dc82cb3cf/frr-k8s-webhook-server/0.log" Dec 05 12:41:23 crc kubenswrapper[4807]: I1205 12:41:23.665715 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-d9c5fbb8c-crqlw_f973bc90-75ad-4998-bbf1-50910dc9a636/manager/0.log" Dec 05 12:41:23 crc kubenswrapper[4807]: I1205 12:41:23.752082 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6ppt7_f7a1e998-d292-49df-88b2-832d0ccb1c83/frr/0.log" Dec 05 12:41:23 crc kubenswrapper[4807]: I1205 12:41:23.754868 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerStarted","Data":"ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c"} Dec 05 12:41:23 crc kubenswrapper[4807]: I1205 12:41:23.774747 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" podStartSLOduration=141.168428918 podStartE2EDuration="2m25.77472938s" podCreationTimestamp="2025-12-05 12:38:58 +0000 UTC" firstStartedPulling="2025-12-05 12:38:59.26861853 +0000 UTC m=+1968.762481799" lastFinishedPulling="2025-12-05 12:39:03.874918982 +0000 UTC m=+1973.368782261" observedRunningTime="2025-12-05 12:39:04.659924456 +0000 UTC m=+1974.153787735" watchObservedRunningTime="2025-12-05 12:41:23.77472938 +0000 UTC m=+2113.268592649" Dec 05 12:41:23 crc kubenswrapper[4807]: I1205 12:41:23.878950 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-745b94f7cd-2h78c_64233fc3-af41-47db-8676-ba182b32e5f7/webhook-server/0.log" Dec 05 12:41:24 crc kubenswrapper[4807]: I1205 12:41:24.018127 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5tvpp_8532b350-fb6d-442c-b388-9a840576a99b/kube-rbac-proxy/0.log" Dec 05 12:41:24 crc kubenswrapper[4807]: I1205 12:41:24.303302 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5tvpp_8532b350-fb6d-442c-b388-9a840576a99b/speaker/0.log" Dec 05 12:41:30 crc kubenswrapper[4807]: I1205 12:41:30.609196 4807 scope.go:117] "RemoveContainer" containerID="2c50c3f64b92e30a1cb34b33dabc5d960cb04ceccd85e1ebc40cff7e1cf7bea4" Dec 05 12:41:30 crc kubenswrapper[4807]: I1205 12:41:30.634124 4807 scope.go:117] "RemoveContainer" containerID="1576e052285f75fd620fba5a71d93132c29733de072e27420457f1ec7738ca95" Dec 05 12:41:30 crc kubenswrapper[4807]: I1205 12:41:30.682490 4807 scope.go:117] "RemoveContainer" containerID="d7d356063957afa6a96ad91c0b931778ba3ab806a4274ed80b5f482b721a0dfd" Dec 05 12:41:30 crc kubenswrapper[4807]: I1205 12:41:30.707845 4807 scope.go:117] "RemoveContainer" containerID="c38e3e7a05d0f6dc7d2e811b8fd0f33138de0bba3977593802d46d8d1e58971e" Dec 05 12:41:30 crc kubenswrapper[4807]: I1205 12:41:30.755366 4807 scope.go:117] "RemoveContainer" containerID="364c39668fdccbfb9483f968615480a26039292004744cd64cf4164299b11a29" Dec 05 12:41:30 crc kubenswrapper[4807]: I1205 12:41:30.780486 4807 scope.go:117] "RemoveContainer" containerID="717435056d17930b4e186aadf19c3235bb2adcbe0eea50456a989c8b797dfb0b" Dec 05 12:41:49 crc kubenswrapper[4807]: I1205 12:41:49.581297 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_4f4aba46-1f95-43ed-8346-57da458a508f/init-config-reloader/0.log" Dec 05 12:41:49 crc kubenswrapper[4807]: I1205 12:41:49.829957 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_4f4aba46-1f95-43ed-8346-57da458a508f/init-config-reloader/0.log" Dec 05 12:41:49 crc kubenswrapper[4807]: I1205 12:41:49.839272 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_4f4aba46-1f95-43ed-8346-57da458a508f/config-reloader/0.log" Dec 05 12:41:49 crc kubenswrapper[4807]: I1205 12:41:49.842046 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_4f4aba46-1f95-43ed-8346-57da458a508f/alertmanager/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.035542 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_db846f0a-96f1-4a17-bdcc-b65bb02094b2/ceilometer-notification-agent/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.106080 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_db846f0a-96f1-4a17-bdcc-b65bb02094b2/ceilometer-central-agent/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.114355 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_db846f0a-96f1-4a17-bdcc-b65bb02094b2/proxy-httpd/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.155141 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_db846f0a-96f1-4a17-bdcc-b65bb02094b2/sg-core/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.387562 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_keystone-5c74bb55d6-5dfz2_34385b79-da95-4009-82a1-6e927d2f6d5b/keystone-api/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.414420 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_keystone-bootstrap-fk576_15cb5d7c-92f9-4a3c-b279-a1b36844c133/keystone-bootstrap/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.593089 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_kube-state-metrics-0_503ab53a-ec37-4100-9021-a3e7836da8ae/kube-state-metrics/0.log" Dec 05 12:41:50 crc kubenswrapper[4807]: I1205 12:41:50.856345 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstack-galera-0_d3244125-122c-4cb6-b673-acc9bdb7e96f/mysql-bootstrap/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.159275 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstack-galera-0_d3244125-122c-4cb6-b673-acc9bdb7e96f/galera/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.229245 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstack-galera-0_d3244125-122c-4cb6-b673-acc9bdb7e96f/mysql-bootstrap/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.466429 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstackclient_118370dc-bda4-46d4-a447-c3022ae95767/openstackclient/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.555761 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_6f4eba60-f9fe-4d26-b6f6-76a326b3e95d/init-config-reloader/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.739477 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_6f4eba60-f9fe-4d26-b6f6-76a326b3e95d/init-config-reloader/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.796390 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_6f4eba60-f9fe-4d26-b6f6-76a326b3e95d/config-reloader/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.952019 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_6f4eba60-f9fe-4d26-b6f6-76a326b3e95d/thanos-sidecar/0.log" Dec 05 12:41:51 crc kubenswrapper[4807]: I1205 12:41:51.964026 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_6f4eba60-f9fe-4d26-b6f6-76a326b3e95d/prometheus/0.log" Dec 05 12:41:52 crc kubenswrapper[4807]: I1205 12:41:52.249032 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-notifications-server-0_3c5fe773-7fd0-4c59-b0c9-1391e8944c1e/setup-container/0.log" Dec 05 12:41:52 crc kubenswrapper[4807]: I1205 12:41:52.706327 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-notifications-server-0_3c5fe773-7fd0-4c59-b0c9-1391e8944c1e/setup-container/0.log" Dec 05 12:41:52 crc kubenswrapper[4807]: I1205 12:41:52.909253 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-notifications-server-0_3c5fe773-7fd0-4c59-b0c9-1391e8944c1e/rabbitmq/0.log" Dec 05 12:41:53 crc kubenswrapper[4807]: I1205 12:41:53.014969 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-server-0_74aad9a5-73e5-45a4-9087-c245511adcf0/setup-container/0.log" Dec 05 12:41:53 crc kubenswrapper[4807]: I1205 12:41:53.149361 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-server-0_74aad9a5-73e5-45a4-9087-c245511adcf0/setup-container/0.log" Dec 05 12:41:53 crc kubenswrapper[4807]: I1205 12:41:53.285396 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-server-0_74aad9a5-73e5-45a4-9087-c245511adcf0/rabbitmq/0.log" Dec 05 12:42:00 crc kubenswrapper[4807]: I1205 12:42:00.130261 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_memcached-0_c483e884-62c1-43bf-9428-5b707b5a2cad/memcached/0.log" Dec 05 12:42:10 crc kubenswrapper[4807]: I1205 12:42:10.707773 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff_eddd7b9a-1c09-4b02-bc51-67cb07107100/util/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.022372 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff_eddd7b9a-1c09-4b02-bc51-67cb07107100/util/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.202092 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff_eddd7b9a-1c09-4b02-bc51-67cb07107100/pull/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.243766 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff_eddd7b9a-1c09-4b02-bc51-67cb07107100/pull/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.495984 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff_eddd7b9a-1c09-4b02-bc51-67cb07107100/pull/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.498975 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff_eddd7b9a-1c09-4b02-bc51-67cb07107100/util/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.524892 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a2h8ff_eddd7b9a-1c09-4b02-bc51-67cb07107100/extract/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.704486 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm_471da076-74b7-4d22-b5ac-67d760a8ee53/util/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.928299 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm_471da076-74b7-4d22-b5ac-67d760a8ee53/pull/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.949345 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm_471da076-74b7-4d22-b5ac-67d760a8ee53/util/0.log" Dec 05 12:42:11 crc kubenswrapper[4807]: I1205 12:42:11.978659 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm_471da076-74b7-4d22-b5ac-67d760a8ee53/pull/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.126178 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm_471da076-74b7-4d22-b5ac-67d760a8ee53/pull/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.140045 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm_471da076-74b7-4d22-b5ac-67d760a8ee53/util/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.153003 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs9gsm_471da076-74b7-4d22-b5ac-67d760a8ee53/extract/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.332357 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l_33a3d26a-4d53-4044-bbd6-bf6b704243ee/util/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.619272 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l_33a3d26a-4d53-4044-bbd6-bf6b704243ee/pull/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.621670 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l_33a3d26a-4d53-4044-bbd6-bf6b704243ee/util/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.632423 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l_33a3d26a-4d53-4044-bbd6-bf6b704243ee/pull/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.842103 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l_33a3d26a-4d53-4044-bbd6-bf6b704243ee/extract/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.887533 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l_33a3d26a-4d53-4044-bbd6-bf6b704243ee/util/0.log" Dec 05 12:42:12 crc kubenswrapper[4807]: I1205 12:42:12.903154 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vw89l_33a3d26a-4d53-4044-bbd6-bf6b704243ee/pull/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.057625 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8_246b60bf-03b5-4139-a13d-691da402cc5f/util/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.258010 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8_246b60bf-03b5-4139-a13d-691da402cc5f/pull/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.268406 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8_246b60bf-03b5-4139-a13d-691da402cc5f/util/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.298190 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8_246b60bf-03b5-4139-a13d-691da402cc5f/pull/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.433208 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8_246b60bf-03b5-4139-a13d-691da402cc5f/pull/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.447986 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8_246b60bf-03b5-4139-a13d-691da402cc5f/extract/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.481111 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kqxs8_246b60bf-03b5-4139-a13d-691da402cc5f/util/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.593264 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84xwf_aa7d7861-9998-4174-974a-47b939d4c4bf/extract-utilities/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.806093 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84xwf_aa7d7861-9998-4174-974a-47b939d4c4bf/extract-content/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.830006 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84xwf_aa7d7861-9998-4174-974a-47b939d4c4bf/extract-content/0.log" Dec 05 12:42:13 crc kubenswrapper[4807]: I1205 12:42:13.857660 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84xwf_aa7d7861-9998-4174-974a-47b939d4c4bf/extract-utilities/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.017639 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84xwf_aa7d7861-9998-4174-974a-47b939d4c4bf/extract-utilities/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.081802 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84xwf_aa7d7861-9998-4174-974a-47b939d4c4bf/extract-content/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.347791 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v99qd_2b7c31e6-006a-450f-af53-e07c2fc71499/extract-utilities/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.370693 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84xwf_aa7d7861-9998-4174-974a-47b939d4c4bf/registry-server/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.596515 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v99qd_2b7c31e6-006a-450f-af53-e07c2fc71499/extract-utilities/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.633728 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v99qd_2b7c31e6-006a-450f-af53-e07c2fc71499/extract-content/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.638049 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v99qd_2b7c31e6-006a-450f-af53-e07c2fc71499/extract-content/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.895709 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v99qd_2b7c31e6-006a-450f-af53-e07c2fc71499/extract-utilities/0.log" Dec 05 12:42:14 crc kubenswrapper[4807]: I1205 12:42:14.895733 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v99qd_2b7c31e6-006a-450f-af53-e07c2fc71499/extract-content/0.log" Dec 05 12:42:15 crc kubenswrapper[4807]: I1205 12:42:15.317572 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v99qd_2b7c31e6-006a-450f-af53-e07c2fc71499/registry-server/0.log" Dec 05 12:42:15 crc kubenswrapper[4807]: I1205 12:42:15.457195 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-m26dt_47587201-3528-4f92-8e07-09afeedcc8eb/marketplace-operator/0.log" Dec 05 12:42:15 crc kubenswrapper[4807]: I1205 12:42:15.517517 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6f2l5_93c28504-1b38-4615-98f2-529b0bfb088f/extract-utilities/0.log" Dec 05 12:42:15 crc kubenswrapper[4807]: I1205 12:42:15.756761 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6f2l5_93c28504-1b38-4615-98f2-529b0bfb088f/extract-utilities/0.log" Dec 05 12:42:15 crc kubenswrapper[4807]: I1205 12:42:15.785115 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6f2l5_93c28504-1b38-4615-98f2-529b0bfb088f/extract-content/0.log" Dec 05 12:42:15 crc kubenswrapper[4807]: I1205 12:42:15.820620 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6f2l5_93c28504-1b38-4615-98f2-529b0bfb088f/extract-content/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.001099 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6f2l5_93c28504-1b38-4615-98f2-529b0bfb088f/extract-utilities/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.034745 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6f2l5_93c28504-1b38-4615-98f2-529b0bfb088f/extract-content/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.059058 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hmcwq_c5bd9595-697e-4315-a23f-11f90b8a8104/extract-utilities/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.126256 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6f2l5_93c28504-1b38-4615-98f2-529b0bfb088f/registry-server/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.331467 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hmcwq_c5bd9595-697e-4315-a23f-11f90b8a8104/extract-content/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.334786 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hmcwq_c5bd9595-697e-4315-a23f-11f90b8a8104/extract-content/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.357535 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hmcwq_c5bd9595-697e-4315-a23f-11f90b8a8104/extract-utilities/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.504631 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hmcwq_c5bd9595-697e-4315-a23f-11f90b8a8104/extract-content/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.508277 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hmcwq_c5bd9595-697e-4315-a23f-11f90b8a8104/extract-utilities/0.log" Dec 05 12:42:16 crc kubenswrapper[4807]: I1205 12:42:16.866582 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hmcwq_c5bd9595-697e-4315-a23f-11f90b8a8104/registry-server/0.log" Dec 05 12:42:29 crc kubenswrapper[4807]: I1205 12:42:29.725078 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-7pnfk_c216b0c0-8388-42f5-96ec-618c61fa131b/prometheus-operator/0.log" Dec 05 12:42:29 crc kubenswrapper[4807]: I1205 12:42:29.852327 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7fcddd784b-mm25w_aef16c20-caa8-4a1b-8b8a-df70e1588307/prometheus-operator-admission-webhook/0.log" Dec 05 12:42:29 crc kubenswrapper[4807]: I1205 12:42:29.952402 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7fcddd784b-nmhtg_73c61937-cac2-4ad4-abe6-4030f1fcdc0d/prometheus-operator-admission-webhook/0.log" Dec 05 12:42:30 crc kubenswrapper[4807]: I1205 12:42:30.060791 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-fk576"] Dec 05 12:42:30 crc kubenswrapper[4807]: I1205 12:42:30.073025 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-fk576"] Dec 05 12:42:30 crc kubenswrapper[4807]: I1205 12:42:30.095658 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-fvbbz_d571b11d-03a4-4879-9bf7-956322d80939/operator/0.log" Dec 05 12:42:30 crc kubenswrapper[4807]: I1205 12:42:30.212226 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-t82lg_d9471f89-372d-4267-aad8-a850cbee6d37/observability-ui-dashboards/0.log" Dec 05 12:42:30 crc kubenswrapper[4807]: I1205 12:42:30.342681 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-hd7g8_78a728e0-7f03-44b3-8a15-bf94733000b2/perses-operator/0.log" Dec 05 12:42:30 crc kubenswrapper[4807]: I1205 12:42:30.909928 4807 scope.go:117] "RemoveContainer" containerID="ff9557c277c332d4bf78a3efd6fb9d0c81a4fac8cccc80a5a8cc997c94538cf6" Dec 05 12:42:30 crc kubenswrapper[4807]: I1205 12:42:30.930952 4807 scope.go:117] "RemoveContainer" containerID="2440865f212d0a177000e6a29ebc554fb0a75168bca253fb04e424489c5a4e24" Dec 05 12:42:31 crc kubenswrapper[4807]: I1205 12:42:31.245056 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15cb5d7c-92f9-4a3c-b279-a1b36844c133" path="/var/lib/kubelet/pods/15cb5d7c-92f9-4a3c-b279-a1b36844c133/volumes" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.074203 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ntw8q"] Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.076450 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.093137 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntw8q"] Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.201692 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjf5b\" (UniqueName: \"kubernetes.io/projected/8ede24e4-7f82-40fd-add7-19ef586179cc-kube-api-access-qjf5b\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.201771 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-catalog-content\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.201843 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-utilities\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.274560 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2f9xc"] Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.276572 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.286875 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2f9xc"] Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.302869 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjf5b\" (UniqueName: \"kubernetes.io/projected/8ede24e4-7f82-40fd-add7-19ef586179cc-kube-api-access-qjf5b\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.302940 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-catalog-content\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.302981 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-utilities\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.303502 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-catalog-content\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.303586 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-utilities\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.339801 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjf5b\" (UniqueName: \"kubernetes.io/projected/8ede24e4-7f82-40fd-add7-19ef586179cc-kube-api-access-qjf5b\") pod \"community-operators-ntw8q\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.392996 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.404207 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-utilities\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.404335 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgmcc\" (UniqueName: \"kubernetes.io/projected/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-kube-api-access-tgmcc\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.404365 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-catalog-content\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.505754 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-utilities\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.505869 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgmcc\" (UniqueName: \"kubernetes.io/projected/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-kube-api-access-tgmcc\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.505897 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-catalog-content\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.506541 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-utilities\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.506689 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-catalog-content\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.532632 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgmcc\" (UniqueName: \"kubernetes.io/projected/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-kube-api-access-tgmcc\") pod \"certified-operators-2f9xc\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.596345 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:17 crc kubenswrapper[4807]: I1205 12:43:17.929915 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntw8q"] Dec 05 12:43:18 crc kubenswrapper[4807]: I1205 12:43:18.119515 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2f9xc"] Dec 05 12:43:18 crc kubenswrapper[4807]: I1205 12:43:18.740704 4807 generic.go:334] "Generic (PLEG): container finished" podID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerID="6462e05d88a27793051c99ab7b03d6a9fe0cd8dd790bb6bab63f78a8061abd0a" exitCode=0 Dec 05 12:43:18 crc kubenswrapper[4807]: I1205 12:43:18.740963 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntw8q" event={"ID":"8ede24e4-7f82-40fd-add7-19ef586179cc","Type":"ContainerDied","Data":"6462e05d88a27793051c99ab7b03d6a9fe0cd8dd790bb6bab63f78a8061abd0a"} Dec 05 12:43:18 crc kubenswrapper[4807]: I1205 12:43:18.741121 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntw8q" event={"ID":"8ede24e4-7f82-40fd-add7-19ef586179cc","Type":"ContainerStarted","Data":"a5b605c648b018a6528f6cf0a0d5c08cec45e99f1987f49de93b1d1f42fa7832"} Dec 05 12:43:18 crc kubenswrapper[4807]: I1205 12:43:18.743252 4807 generic.go:334] "Generic (PLEG): container finished" podID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerID="263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c" exitCode=0 Dec 05 12:43:18 crc kubenswrapper[4807]: I1205 12:43:18.743276 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2f9xc" event={"ID":"6581c3a3-b0a0-4b06-bfd9-3f122a63528b","Type":"ContainerDied","Data":"263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c"} Dec 05 12:43:18 crc kubenswrapper[4807]: I1205 12:43:18.743293 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2f9xc" event={"ID":"6581c3a3-b0a0-4b06-bfd9-3f122a63528b","Type":"ContainerStarted","Data":"5a942d2557835e3ad3c4c80462c144843801a1a682d75337af6a507ac59358c3"} Dec 05 12:43:20 crc kubenswrapper[4807]: I1205 12:43:20.761108 4807 generic.go:334] "Generic (PLEG): container finished" podID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerID="24361294c654eddf2d1866dee5ef857191052fb8564e9d56078eca82194b4488" exitCode=0 Dec 05 12:43:20 crc kubenswrapper[4807]: I1205 12:43:20.761214 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntw8q" event={"ID":"8ede24e4-7f82-40fd-add7-19ef586179cc","Type":"ContainerDied","Data":"24361294c654eddf2d1866dee5ef857191052fb8564e9d56078eca82194b4488"} Dec 05 12:43:20 crc kubenswrapper[4807]: I1205 12:43:20.764244 4807 generic.go:334] "Generic (PLEG): container finished" podID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerID="7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94" exitCode=0 Dec 05 12:43:20 crc kubenswrapper[4807]: I1205 12:43:20.764295 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2f9xc" event={"ID":"6581c3a3-b0a0-4b06-bfd9-3f122a63528b","Type":"ContainerDied","Data":"7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94"} Dec 05 12:43:21 crc kubenswrapper[4807]: I1205 12:43:21.777623 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntw8q" event={"ID":"8ede24e4-7f82-40fd-add7-19ef586179cc","Type":"ContainerStarted","Data":"5d38a8b1b235cb3e099ca2b72ce2fea119d2766e8a4dbd049fb0142452d39e12"} Dec 05 12:43:21 crc kubenswrapper[4807]: I1205 12:43:21.781656 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2f9xc" event={"ID":"6581c3a3-b0a0-4b06-bfd9-3f122a63528b","Type":"ContainerStarted","Data":"efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4"} Dec 05 12:43:21 crc kubenswrapper[4807]: I1205 12:43:21.799151 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ntw8q" podStartSLOduration=2.295058994 podStartE2EDuration="4.799134434s" podCreationTimestamp="2025-12-05 12:43:17 +0000 UTC" firstStartedPulling="2025-12-05 12:43:18.744196513 +0000 UTC m=+2228.238059782" lastFinishedPulling="2025-12-05 12:43:21.248271953 +0000 UTC m=+2230.742135222" observedRunningTime="2025-12-05 12:43:21.796984751 +0000 UTC m=+2231.290848020" watchObservedRunningTime="2025-12-05 12:43:21.799134434 +0000 UTC m=+2231.292997703" Dec 05 12:43:21 crc kubenswrapper[4807]: I1205 12:43:21.821111 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2f9xc" podStartSLOduration=2.403103814 podStartE2EDuration="4.821089118s" podCreationTimestamp="2025-12-05 12:43:17 +0000 UTC" firstStartedPulling="2025-12-05 12:43:18.744808408 +0000 UTC m=+2228.238671677" lastFinishedPulling="2025-12-05 12:43:21.162793702 +0000 UTC m=+2230.656656981" observedRunningTime="2025-12-05 12:43:21.820542095 +0000 UTC m=+2231.314405374" watchObservedRunningTime="2025-12-05 12:43:21.821089118 +0000 UTC m=+2231.314952387" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.396072 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.397742 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.449630 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.597637 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.597676 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.643827 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.874114 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:27 crc kubenswrapper[4807]: I1205 12:43:27.876430 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:29 crc kubenswrapper[4807]: I1205 12:43:29.259533 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2f9xc"] Dec 05 12:43:29 crc kubenswrapper[4807]: I1205 12:43:29.849308 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2f9xc" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="registry-server" containerID="cri-o://efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4" gracePeriod=2 Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.296690 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.439709 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-utilities\") pod \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.439788 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgmcc\" (UniqueName: \"kubernetes.io/projected/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-kube-api-access-tgmcc\") pod \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.439815 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-catalog-content\") pod \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\" (UID: \"6581c3a3-b0a0-4b06-bfd9-3f122a63528b\") " Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.440398 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-utilities" (OuterVolumeSpecName: "utilities") pod "6581c3a3-b0a0-4b06-bfd9-3f122a63528b" (UID: "6581c3a3-b0a0-4b06-bfd9-3f122a63528b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.452197 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-kube-api-access-tgmcc" (OuterVolumeSpecName: "kube-api-access-tgmcc") pod "6581c3a3-b0a0-4b06-bfd9-3f122a63528b" (UID: "6581c3a3-b0a0-4b06-bfd9-3f122a63528b"). InnerVolumeSpecName "kube-api-access-tgmcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.541656 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.541688 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgmcc\" (UniqueName: \"kubernetes.io/projected/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-kube-api-access-tgmcc\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.859413 4807 generic.go:334] "Generic (PLEG): container finished" podID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerID="efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4" exitCode=0 Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.859458 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2f9xc" event={"ID":"6581c3a3-b0a0-4b06-bfd9-3f122a63528b","Type":"ContainerDied","Data":"efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4"} Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.859911 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2f9xc" event={"ID":"6581c3a3-b0a0-4b06-bfd9-3f122a63528b","Type":"ContainerDied","Data":"5a942d2557835e3ad3c4c80462c144843801a1a682d75337af6a507ac59358c3"} Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.859575 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2f9xc" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.859981 4807 scope.go:117] "RemoveContainer" containerID="efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.879551 4807 scope.go:117] "RemoveContainer" containerID="7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.906372 4807 scope.go:117] "RemoveContainer" containerID="263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.941411 4807 scope.go:117] "RemoveContainer" containerID="efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4" Dec 05 12:43:30 crc kubenswrapper[4807]: E1205 12:43:30.941822 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4\": container with ID starting with efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4 not found: ID does not exist" containerID="efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.941863 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4"} err="failed to get container status \"efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4\": rpc error: code = NotFound desc = could not find container \"efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4\": container with ID starting with efbd9879091c2304ee8402ca3c83a2c442f3d42ff1340c34c6a4db2c694a4eb4 not found: ID does not exist" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.941893 4807 scope.go:117] "RemoveContainer" containerID="7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94" Dec 05 12:43:30 crc kubenswrapper[4807]: E1205 12:43:30.942270 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94\": container with ID starting with 7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94 not found: ID does not exist" containerID="7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.942292 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94"} err="failed to get container status \"7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94\": rpc error: code = NotFound desc = could not find container \"7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94\": container with ID starting with 7ef7304b5c04c16d2650a63db0635f27c2d879ff4a91b389e5f7695adab58a94 not found: ID does not exist" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.942307 4807 scope.go:117] "RemoveContainer" containerID="263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c" Dec 05 12:43:30 crc kubenswrapper[4807]: E1205 12:43:30.947752 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c\": container with ID starting with 263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c not found: ID does not exist" containerID="263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.947795 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c"} err="failed to get container status \"263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c\": rpc error: code = NotFound desc = could not find container \"263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c\": container with ID starting with 263fc4ee04b7abbbb6868f75862e5705ceae593b038b47a2f65b033c307bf40c not found: ID does not exist" Dec 05 12:43:30 crc kubenswrapper[4807]: I1205 12:43:30.949839 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6581c3a3-b0a0-4b06-bfd9-3f122a63528b" (UID: "6581c3a3-b0a0-4b06-bfd9-3f122a63528b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.022675 4807 scope.go:117] "RemoveContainer" containerID="fdc8938a67fc44474596e2b1498b97f89d2bd91f7de957188d5834956d170b3c" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.049614 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6581c3a3-b0a0-4b06-bfd9-3f122a63528b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.063188 4807 scope.go:117] "RemoveContainer" containerID="9fbd2ecb81731c6c95659d8d3b5e7ef166d645a9911815a165f17e00793137c8" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.084960 4807 scope.go:117] "RemoveContainer" containerID="d030f02cd0f90cf768c4d2564f212dd593850c95817efe9c458466124177ca37" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.126251 4807 scope.go:117] "RemoveContainer" containerID="dc3ff126ea0ecf2857481fe9694669ba2cd1d87a737b4ac189f91d2ebadfac99" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.169946 4807 scope.go:117] "RemoveContainer" containerID="7a936439c18f8b9d66bd1647bb08616ea470c1fee1b313845a5a94efeb102f53" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.199776 4807 scope.go:117] "RemoveContainer" containerID="d9f118d16acd2df0531590783830463dbfd4d4d5e6003d5fd7df3399ba195ec7" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.202572 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2f9xc"] Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.210426 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2f9xc"] Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.216540 4807 scope.go:117] "RemoveContainer" containerID="a72af4e002c5d14d59d9b27a24133ce103a8c6fdc779ef88799a428132c69fd0" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.236371 4807 scope.go:117] "RemoveContainer" containerID="926d26b4ff019b32dcedffab80d1bf4283c824b401f711f69ea5d40dbd5be1b0" Dec 05 12:43:31 crc kubenswrapper[4807]: I1205 12:43:31.245851 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" path="/var/lib/kubelet/pods/6581c3a3-b0a0-4b06-bfd9-3f122a63528b/volumes" Dec 05 12:43:34 crc kubenswrapper[4807]: I1205 12:43:34.056643 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntw8q"] Dec 05 12:43:34 crc kubenswrapper[4807]: I1205 12:43:34.058446 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ntw8q" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="registry-server" containerID="cri-o://5d38a8b1b235cb3e099ca2b72ce2fea119d2766e8a4dbd049fb0142452d39e12" gracePeriod=2 Dec 05 12:43:34 crc kubenswrapper[4807]: I1205 12:43:34.896055 4807 generic.go:334] "Generic (PLEG): container finished" podID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerID="5d38a8b1b235cb3e099ca2b72ce2fea119d2766e8a4dbd049fb0142452d39e12" exitCode=0 Dec 05 12:43:34 crc kubenswrapper[4807]: I1205 12:43:34.896105 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntw8q" event={"ID":"8ede24e4-7f82-40fd-add7-19ef586179cc","Type":"ContainerDied","Data":"5d38a8b1b235cb3e099ca2b72ce2fea119d2766e8a4dbd049fb0142452d39e12"} Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.017588 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.105681 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-utilities\") pod \"8ede24e4-7f82-40fd-add7-19ef586179cc\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.105812 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-catalog-content\") pod \"8ede24e4-7f82-40fd-add7-19ef586179cc\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.105877 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjf5b\" (UniqueName: \"kubernetes.io/projected/8ede24e4-7f82-40fd-add7-19ef586179cc-kube-api-access-qjf5b\") pod \"8ede24e4-7f82-40fd-add7-19ef586179cc\" (UID: \"8ede24e4-7f82-40fd-add7-19ef586179cc\") " Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.106825 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-utilities" (OuterVolumeSpecName: "utilities") pod "8ede24e4-7f82-40fd-add7-19ef586179cc" (UID: "8ede24e4-7f82-40fd-add7-19ef586179cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.112443 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ede24e4-7f82-40fd-add7-19ef586179cc-kube-api-access-qjf5b" (OuterVolumeSpecName: "kube-api-access-qjf5b") pod "8ede24e4-7f82-40fd-add7-19ef586179cc" (UID: "8ede24e4-7f82-40fd-add7-19ef586179cc"). InnerVolumeSpecName "kube-api-access-qjf5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.163888 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ede24e4-7f82-40fd-add7-19ef586179cc" (UID: "8ede24e4-7f82-40fd-add7-19ef586179cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.207940 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjf5b\" (UniqueName: \"kubernetes.io/projected/8ede24e4-7f82-40fd-add7-19ef586179cc-kube-api-access-qjf5b\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.207983 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.207995 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ede24e4-7f82-40fd-add7-19ef586179cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.906033 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntw8q" event={"ID":"8ede24e4-7f82-40fd-add7-19ef586179cc","Type":"ContainerDied","Data":"a5b605c648b018a6528f6cf0a0d5c08cec45e99f1987f49de93b1d1f42fa7832"} Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.906093 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntw8q" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.906098 4807 scope.go:117] "RemoveContainer" containerID="5d38a8b1b235cb3e099ca2b72ce2fea119d2766e8a4dbd049fb0142452d39e12" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.929239 4807 scope.go:117] "RemoveContainer" containerID="24361294c654eddf2d1866dee5ef857191052fb8564e9d56078eca82194b4488" Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.941119 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntw8q"] Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.949004 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ntw8q"] Dec 05 12:43:35 crc kubenswrapper[4807]: I1205 12:43:35.955603 4807 scope.go:117] "RemoveContainer" containerID="6462e05d88a27793051c99ab7b03d6a9fe0cd8dd790bb6bab63f78a8061abd0a" Dec 05 12:43:36 crc kubenswrapper[4807]: I1205 12:43:36.918632 4807 generic.go:334] "Generic (PLEG): container finished" podID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerID="0c46ae024a3db1eb2624c1331f1927076c0edf3a4bcbbeaf418722b3b18874fc" exitCode=0 Dec 05 12:43:36 crc kubenswrapper[4807]: I1205 12:43:36.918681 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" event={"ID":"a1da4dc7-4134-414d-8494-bf7bd296d80f","Type":"ContainerDied","Data":"0c46ae024a3db1eb2624c1331f1927076c0edf3a4bcbbeaf418722b3b18874fc"} Dec 05 12:43:36 crc kubenswrapper[4807]: I1205 12:43:36.919344 4807 scope.go:117] "RemoveContainer" containerID="0c46ae024a3db1eb2624c1331f1927076c0edf3a4bcbbeaf418722b3b18874fc" Dec 05 12:43:37 crc kubenswrapper[4807]: I1205 12:43:37.249935 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" path="/var/lib/kubelet/pods/8ede24e4-7f82-40fd-add7-19ef586179cc/volumes" Dec 05 12:43:37 crc kubenswrapper[4807]: I1205 12:43:37.347115 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zm9tk_must-gather-8lb9h_a1da4dc7-4134-414d-8494-bf7bd296d80f/gather/0.log" Dec 05 12:43:44 crc kubenswrapper[4807]: I1205 12:43:44.752056 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zm9tk/must-gather-8lb9h"] Dec 05 12:43:44 crc kubenswrapper[4807]: I1205 12:43:44.752786 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerName="copy" containerID="cri-o://902a9252892d45cd26292d61bc1710505d9a5fc057b91241a6206ac39d089bc8" gracePeriod=2 Dec 05 12:43:44 crc kubenswrapper[4807]: I1205 12:43:44.758208 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zm9tk/must-gather-8lb9h"] Dec 05 12:43:44 crc kubenswrapper[4807]: I1205 12:43:44.988151 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zm9tk_must-gather-8lb9h_a1da4dc7-4134-414d-8494-bf7bd296d80f/copy/0.log" Dec 05 12:43:44 crc kubenswrapper[4807]: I1205 12:43:44.988648 4807 generic.go:334] "Generic (PLEG): container finished" podID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerID="902a9252892d45cd26292d61bc1710505d9a5fc057b91241a6206ac39d089bc8" exitCode=143 Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.127260 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zm9tk_must-gather-8lb9h_a1da4dc7-4134-414d-8494-bf7bd296d80f/copy/0.log" Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.128104 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.179410 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4cfx\" (UniqueName: \"kubernetes.io/projected/a1da4dc7-4134-414d-8494-bf7bd296d80f-kube-api-access-t4cfx\") pod \"a1da4dc7-4134-414d-8494-bf7bd296d80f\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.179538 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a1da4dc7-4134-414d-8494-bf7bd296d80f-must-gather-output\") pod \"a1da4dc7-4134-414d-8494-bf7bd296d80f\" (UID: \"a1da4dc7-4134-414d-8494-bf7bd296d80f\") " Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.184989 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1da4dc7-4134-414d-8494-bf7bd296d80f-kube-api-access-t4cfx" (OuterVolumeSpecName: "kube-api-access-t4cfx") pod "a1da4dc7-4134-414d-8494-bf7bd296d80f" (UID: "a1da4dc7-4134-414d-8494-bf7bd296d80f"). InnerVolumeSpecName "kube-api-access-t4cfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.281904 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1da4dc7-4134-414d-8494-bf7bd296d80f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a1da4dc7-4134-414d-8494-bf7bd296d80f" (UID: "a1da4dc7-4134-414d-8494-bf7bd296d80f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.283012 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4cfx\" (UniqueName: \"kubernetes.io/projected/a1da4dc7-4134-414d-8494-bf7bd296d80f-kube-api-access-t4cfx\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.283032 4807 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a1da4dc7-4134-414d-8494-bf7bd296d80f-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:45 crc kubenswrapper[4807]: I1205 12:43:45.999701 4807 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zm9tk_must-gather-8lb9h_a1da4dc7-4134-414d-8494-bf7bd296d80f/copy/0.log" Dec 05 12:43:46 crc kubenswrapper[4807]: I1205 12:43:46.000156 4807 scope.go:117] "RemoveContainer" containerID="902a9252892d45cd26292d61bc1710505d9a5fc057b91241a6206ac39d089bc8" Dec 05 12:43:46 crc kubenswrapper[4807]: I1205 12:43:46.000321 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zm9tk/must-gather-8lb9h" Dec 05 12:43:46 crc kubenswrapper[4807]: I1205 12:43:46.030246 4807 scope.go:117] "RemoveContainer" containerID="0c46ae024a3db1eb2624c1331f1927076c0edf3a4bcbbeaf418722b3b18874fc" Dec 05 12:43:47 crc kubenswrapper[4807]: I1205 12:43:47.246417 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" path="/var/lib/kubelet/pods/a1da4dc7-4134-414d-8494-bf7bd296d80f/volumes" Dec 05 12:43:52 crc kubenswrapper[4807]: I1205 12:43:52.466292 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:43:52 crc kubenswrapper[4807]: I1205 12:43:52.466866 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:44:22 crc kubenswrapper[4807]: I1205 12:44:22.466316 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:44:22 crc kubenswrapper[4807]: I1205 12:44:22.466871 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:44:31 crc kubenswrapper[4807]: I1205 12:44:31.407566 4807 scope.go:117] "RemoveContainer" containerID="ffa53385c003261df2402b6c6df4263734c55aee847f8445af2f89e163c815df" Dec 05 12:44:31 crc kubenswrapper[4807]: I1205 12:44:31.442718 4807 scope.go:117] "RemoveContainer" containerID="fedbebd8ae752c9fa6d9d23f8d3a5b5c3fb6eccda16303307b316631ec932efe" Dec 05 12:44:31 crc kubenswrapper[4807]: I1205 12:44:31.463294 4807 scope.go:117] "RemoveContainer" containerID="58097ddc93f4c6390ec8aaaef702269da526e77f5cf41f930789e253b18cc765" Dec 05 12:44:52 crc kubenswrapper[4807]: I1205 12:44:52.466365 4807 patch_prober.go:28] interesting pod/machine-config-daemon-kth9r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:44:52 crc kubenswrapper[4807]: I1205 12:44:52.467651 4807 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:44:52 crc kubenswrapper[4807]: I1205 12:44:52.467747 4807 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" Dec 05 12:44:52 crc kubenswrapper[4807]: I1205 12:44:52.469384 4807 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c"} pod="openshift-machine-config-operator/machine-config-daemon-kth9r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:44:52 crc kubenswrapper[4807]: I1205 12:44:52.469472 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerName="machine-config-daemon" containerID="cri-o://ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" gracePeriod=600 Dec 05 12:44:53 crc kubenswrapper[4807]: E1205 12:44:53.099457 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:44:53 crc kubenswrapper[4807]: I1205 12:44:53.543384 4807 generic.go:334] "Generic (PLEG): container finished" podID="a779882f-2b7e-4ae0-addd-686fd4343bb6" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" exitCode=0 Dec 05 12:44:53 crc kubenswrapper[4807]: I1205 12:44:53.543467 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" event={"ID":"a779882f-2b7e-4ae0-addd-686fd4343bb6","Type":"ContainerDied","Data":"ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c"} Dec 05 12:44:53 crc kubenswrapper[4807]: I1205 12:44:53.543712 4807 scope.go:117] "RemoveContainer" containerID="5d1db79e77bd059c1448ecb5528526ada5577ec3cff58469fb098ce90654d049" Dec 05 12:44:53 crc kubenswrapper[4807]: I1205 12:44:53.544246 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:44:53 crc kubenswrapper[4807]: E1205 12:44:53.544489 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.152486 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb"] Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153484 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="registry-server" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153501 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="registry-server" Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153515 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerName="copy" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153522 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerName="copy" Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153556 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="registry-server" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153564 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="registry-server" Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153594 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="extract-content" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153602 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="extract-content" Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153615 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="extract-utilities" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153622 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="extract-utilities" Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153635 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerName="gather" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153641 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerName="gather" Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153661 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="extract-utilities" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153668 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="extract-utilities" Dec 05 12:45:00 crc kubenswrapper[4807]: E1205 12:45:00.153681 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="extract-content" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153688 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="extract-content" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153866 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerName="gather" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153882 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1da4dc7-4134-414d-8494-bf7bd296d80f" containerName="copy" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153895 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="6581c3a3-b0a0-4b06-bfd9-3f122a63528b" containerName="registry-server" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.153918 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ede24e4-7f82-40fd-add7-19ef586179cc" containerName="registry-server" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.154633 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.157187 4807 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.159612 4807 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.192878 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb"] Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.273996 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n75g2\" (UniqueName: \"kubernetes.io/projected/95ea4208-dc23-4666-be87-e643f0a2f76b-kube-api-access-n75g2\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.274340 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95ea4208-dc23-4666-be87-e643f0a2f76b-config-volume\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.274374 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95ea4208-dc23-4666-be87-e643f0a2f76b-secret-volume\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.376517 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n75g2\" (UniqueName: \"kubernetes.io/projected/95ea4208-dc23-4666-be87-e643f0a2f76b-kube-api-access-n75g2\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.376617 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95ea4208-dc23-4666-be87-e643f0a2f76b-config-volume\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.376673 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95ea4208-dc23-4666-be87-e643f0a2f76b-secret-volume\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.378704 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95ea4208-dc23-4666-be87-e643f0a2f76b-config-volume\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.393923 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95ea4208-dc23-4666-be87-e643f0a2f76b-secret-volume\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.393961 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n75g2\" (UniqueName: \"kubernetes.io/projected/95ea4208-dc23-4666-be87-e643f0a2f76b-kube-api-access-n75g2\") pod \"collect-profiles-29415645-v6dbb\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.483177 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:00 crc kubenswrapper[4807]: I1205 12:45:00.901827 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb"] Dec 05 12:45:01 crc kubenswrapper[4807]: I1205 12:45:01.613638 4807 generic.go:334] "Generic (PLEG): container finished" podID="95ea4208-dc23-4666-be87-e643f0a2f76b" containerID="4ed16e2403813880217b7b47856406b3d4fe04d9388952947a149e9ea32b55f8" exitCode=0 Dec 05 12:45:01 crc kubenswrapper[4807]: I1205 12:45:01.613921 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" event={"ID":"95ea4208-dc23-4666-be87-e643f0a2f76b","Type":"ContainerDied","Data":"4ed16e2403813880217b7b47856406b3d4fe04d9388952947a149e9ea32b55f8"} Dec 05 12:45:01 crc kubenswrapper[4807]: I1205 12:45:01.613949 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" event={"ID":"95ea4208-dc23-4666-be87-e643f0a2f76b","Type":"ContainerStarted","Data":"225dfc4189a0e0d9483718fe9c48f123ba388bb6e9e990dac279c1ec97f9ab9c"} Dec 05 12:45:02 crc kubenswrapper[4807]: I1205 12:45:02.963586 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.018809 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95ea4208-dc23-4666-be87-e643f0a2f76b-secret-volume\") pod \"95ea4208-dc23-4666-be87-e643f0a2f76b\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.018850 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n75g2\" (UniqueName: \"kubernetes.io/projected/95ea4208-dc23-4666-be87-e643f0a2f76b-kube-api-access-n75g2\") pod \"95ea4208-dc23-4666-be87-e643f0a2f76b\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.018912 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95ea4208-dc23-4666-be87-e643f0a2f76b-config-volume\") pod \"95ea4208-dc23-4666-be87-e643f0a2f76b\" (UID: \"95ea4208-dc23-4666-be87-e643f0a2f76b\") " Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.019613 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95ea4208-dc23-4666-be87-e643f0a2f76b-config-volume" (OuterVolumeSpecName: "config-volume") pod "95ea4208-dc23-4666-be87-e643f0a2f76b" (UID: "95ea4208-dc23-4666-be87-e643f0a2f76b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.025659 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ea4208-dc23-4666-be87-e643f0a2f76b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "95ea4208-dc23-4666-be87-e643f0a2f76b" (UID: "95ea4208-dc23-4666-be87-e643f0a2f76b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.025700 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ea4208-dc23-4666-be87-e643f0a2f76b-kube-api-access-n75g2" (OuterVolumeSpecName: "kube-api-access-n75g2") pod "95ea4208-dc23-4666-be87-e643f0a2f76b" (UID: "95ea4208-dc23-4666-be87-e643f0a2f76b"). InnerVolumeSpecName "kube-api-access-n75g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.121037 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n75g2\" (UniqueName: \"kubernetes.io/projected/95ea4208-dc23-4666-be87-e643f0a2f76b-kube-api-access-n75g2\") on node \"crc\" DevicePath \"\"" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.121075 4807 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95ea4208-dc23-4666-be87-e643f0a2f76b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.121084 4807 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95ea4208-dc23-4666-be87-e643f0a2f76b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.631375 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" event={"ID":"95ea4208-dc23-4666-be87-e643f0a2f76b","Type":"ContainerDied","Data":"225dfc4189a0e0d9483718fe9c48f123ba388bb6e9e990dac279c1ec97f9ab9c"} Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.631415 4807 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="225dfc4189a0e0d9483718fe9c48f123ba388bb6e9e990dac279c1ec97f9ab9c" Dec 05 12:45:03 crc kubenswrapper[4807]: I1205 12:45:03.631441 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-v6dbb" Dec 05 12:45:04 crc kubenswrapper[4807]: I1205 12:45:04.045025 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx"] Dec 05 12:45:04 crc kubenswrapper[4807]: I1205 12:45:04.055214 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-xwfzx"] Dec 05 12:45:05 crc kubenswrapper[4807]: I1205 12:45:05.245771 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e849f7a7-467d-4124-a342-c8b842e28790" path="/var/lib/kubelet/pods/e849f7a7-467d-4124-a342-c8b842e28790/volumes" Dec 05 12:45:07 crc kubenswrapper[4807]: I1205 12:45:07.239918 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:45:07 crc kubenswrapper[4807]: E1205 12:45:07.240963 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:45:22 crc kubenswrapper[4807]: I1205 12:45:22.235834 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:45:22 crc kubenswrapper[4807]: E1205 12:45:22.236455 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:45:31 crc kubenswrapper[4807]: I1205 12:45:31.611332 4807 scope.go:117] "RemoveContainer" containerID="c3ad01668fff5546c385b6313f72931ffeac99b301c41e4f2a52029b1f138d40" Dec 05 12:45:35 crc kubenswrapper[4807]: I1205 12:45:35.236233 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:45:35 crc kubenswrapper[4807]: E1205 12:45:35.237066 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:45:48 crc kubenswrapper[4807]: I1205 12:45:48.235332 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:45:48 crc kubenswrapper[4807]: E1205 12:45:48.236215 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:46:02 crc kubenswrapper[4807]: I1205 12:46:02.236219 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:46:02 crc kubenswrapper[4807]: E1205 12:46:02.237256 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:46:17 crc kubenswrapper[4807]: I1205 12:46:17.235659 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:46:17 crc kubenswrapper[4807]: E1205 12:46:17.237564 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:46:31 crc kubenswrapper[4807]: I1205 12:46:31.242332 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:46:31 crc kubenswrapper[4807]: E1205 12:46:31.243182 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:46:43 crc kubenswrapper[4807]: I1205 12:46:43.235120 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:46:43 crc kubenswrapper[4807]: E1205 12:46:43.235969 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:46:57 crc kubenswrapper[4807]: I1205 12:46:57.236319 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:46:57 crc kubenswrapper[4807]: E1205 12:46:57.237015 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:47:08 crc kubenswrapper[4807]: I1205 12:47:08.236471 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:47:08 crc kubenswrapper[4807]: E1205 12:47:08.237465 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:47:23 crc kubenswrapper[4807]: I1205 12:47:23.235391 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:47:23 crc kubenswrapper[4807]: E1205 12:47:23.236147 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:47:37 crc kubenswrapper[4807]: I1205 12:47:37.235237 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:47:37 crc kubenswrapper[4807]: E1205 12:47:37.236055 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:47:45 crc kubenswrapper[4807]: I1205 12:47:45.897337 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hbwx4"] Dec 05 12:47:45 crc kubenswrapper[4807]: E1205 12:47:45.898479 4807 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ea4208-dc23-4666-be87-e643f0a2f76b" containerName="collect-profiles" Dec 05 12:47:45 crc kubenswrapper[4807]: I1205 12:47:45.898497 4807 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ea4208-dc23-4666-be87-e643f0a2f76b" containerName="collect-profiles" Dec 05 12:47:45 crc kubenswrapper[4807]: I1205 12:47:45.898775 4807 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ea4208-dc23-4666-be87-e643f0a2f76b" containerName="collect-profiles" Dec 05 12:47:45 crc kubenswrapper[4807]: I1205 12:47:45.906424 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:45 crc kubenswrapper[4807]: I1205 12:47:45.909925 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbwx4"] Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.015380 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-utilities\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.015817 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q4hz\" (UniqueName: \"kubernetes.io/projected/b9b0418b-c966-4dc4-9188-6201b308eaa8-kube-api-access-5q4hz\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.015854 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-catalog-content\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.117066 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-utilities\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.117158 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q4hz\" (UniqueName: \"kubernetes.io/projected/b9b0418b-c966-4dc4-9188-6201b308eaa8-kube-api-access-5q4hz\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.117197 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-catalog-content\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.117905 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-utilities\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.117958 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-catalog-content\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.135151 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q4hz\" (UniqueName: \"kubernetes.io/projected/b9b0418b-c966-4dc4-9188-6201b308eaa8-kube-api-access-5q4hz\") pod \"redhat-operators-hbwx4\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.226433 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:46 crc kubenswrapper[4807]: I1205 12:47:46.706687 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbwx4"] Dec 05 12:47:47 crc kubenswrapper[4807]: I1205 12:47:47.046152 4807 generic.go:334] "Generic (PLEG): container finished" podID="b9b0418b-c966-4dc4-9188-6201b308eaa8" containerID="dee7f8877fb475a9e0383ea936d6907b5602ed8707beb04ef07793a9044a753b" exitCode=0 Dec 05 12:47:47 crc kubenswrapper[4807]: I1205 12:47:47.046251 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbwx4" event={"ID":"b9b0418b-c966-4dc4-9188-6201b308eaa8","Type":"ContainerDied","Data":"dee7f8877fb475a9e0383ea936d6907b5602ed8707beb04ef07793a9044a753b"} Dec 05 12:47:47 crc kubenswrapper[4807]: I1205 12:47:47.047460 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbwx4" event={"ID":"b9b0418b-c966-4dc4-9188-6201b308eaa8","Type":"ContainerStarted","Data":"e0ae3edd5d384542978226ef3f552c59e20397f6d93401e137cf2605e8e6c583"} Dec 05 12:47:47 crc kubenswrapper[4807]: I1205 12:47:47.048873 4807 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:47:48 crc kubenswrapper[4807]: I1205 12:47:48.058882 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbwx4" event={"ID":"b9b0418b-c966-4dc4-9188-6201b308eaa8","Type":"ContainerStarted","Data":"b5f5d796d658100e98481e1c2f487ebab1dad245873b44db488c78b7ce33481e"} Dec 05 12:47:49 crc kubenswrapper[4807]: I1205 12:47:49.067844 4807 generic.go:334] "Generic (PLEG): container finished" podID="b9b0418b-c966-4dc4-9188-6201b308eaa8" containerID="b5f5d796d658100e98481e1c2f487ebab1dad245873b44db488c78b7ce33481e" exitCode=0 Dec 05 12:47:49 crc kubenswrapper[4807]: I1205 12:47:49.068009 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbwx4" event={"ID":"b9b0418b-c966-4dc4-9188-6201b308eaa8","Type":"ContainerDied","Data":"b5f5d796d658100e98481e1c2f487ebab1dad245873b44db488c78b7ce33481e"} Dec 05 12:47:49 crc kubenswrapper[4807]: I1205 12:47:49.235996 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:47:49 crc kubenswrapper[4807]: E1205 12:47:49.236250 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.077161 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbwx4" event={"ID":"b9b0418b-c966-4dc4-9188-6201b308eaa8","Type":"ContainerStarted","Data":"fd96faf251d8dfc69031c1191069fc0009dd20fc042584fca5cc80b7798e0a23"} Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.099447 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hbwx4" podStartSLOduration=2.658219153 podStartE2EDuration="5.099430492s" podCreationTimestamp="2025-12-05 12:47:45 +0000 UTC" firstStartedPulling="2025-12-05 12:47:47.048668017 +0000 UTC m=+2496.542531286" lastFinishedPulling="2025-12-05 12:47:49.489879356 +0000 UTC m=+2498.983742625" observedRunningTime="2025-12-05 12:47:50.097294391 +0000 UTC m=+2499.591157670" watchObservedRunningTime="2025-12-05 12:47:50.099430492 +0000 UTC m=+2499.593293761" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.289939 4807 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qp6fn"] Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.293299 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.301465 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qp6fn"] Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.396450 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h87q\" (UniqueName: \"kubernetes.io/projected/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-kube-api-access-9h87q\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.396557 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-catalog-content\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.396592 4807 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-utilities\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.497955 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-catalog-content\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.498255 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-utilities\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.498348 4807 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h87q\" (UniqueName: \"kubernetes.io/projected/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-kube-api-access-9h87q\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.498552 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-utilities\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.498753 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-catalog-content\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.519760 4807 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h87q\" (UniqueName: \"kubernetes.io/projected/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-kube-api-access-9h87q\") pod \"redhat-marketplace-qp6fn\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:50 crc kubenswrapper[4807]: I1205 12:47:50.617610 4807 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:47:51 crc kubenswrapper[4807]: I1205 12:47:51.111617 4807 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qp6fn"] Dec 05 12:47:52 crc kubenswrapper[4807]: I1205 12:47:52.097881 4807 generic.go:334] "Generic (PLEG): container finished" podID="a0a51e52-ddc5-43d2-afc0-4cd69c24477a" containerID="41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b" exitCode=0 Dec 05 12:47:52 crc kubenswrapper[4807]: I1205 12:47:52.098424 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qp6fn" event={"ID":"a0a51e52-ddc5-43d2-afc0-4cd69c24477a","Type":"ContainerDied","Data":"41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b"} Dec 05 12:47:52 crc kubenswrapper[4807]: I1205 12:47:52.098546 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qp6fn" event={"ID":"a0a51e52-ddc5-43d2-afc0-4cd69c24477a","Type":"ContainerStarted","Data":"8e636f497c3ea13c491258f82924ae7e8d9b33e8fc8fe90e65e39fb122b8a8ff"} Dec 05 12:47:54 crc kubenswrapper[4807]: I1205 12:47:54.115972 4807 generic.go:334] "Generic (PLEG): container finished" podID="a0a51e52-ddc5-43d2-afc0-4cd69c24477a" containerID="2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7" exitCode=0 Dec 05 12:47:54 crc kubenswrapper[4807]: I1205 12:47:54.116020 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qp6fn" event={"ID":"a0a51e52-ddc5-43d2-afc0-4cd69c24477a","Type":"ContainerDied","Data":"2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7"} Dec 05 12:47:55 crc kubenswrapper[4807]: I1205 12:47:55.126055 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qp6fn" event={"ID":"a0a51e52-ddc5-43d2-afc0-4cd69c24477a","Type":"ContainerStarted","Data":"2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58"} Dec 05 12:47:55 crc kubenswrapper[4807]: I1205 12:47:55.148091 4807 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qp6fn" podStartSLOduration=2.7685029439999997 podStartE2EDuration="5.148072954s" podCreationTimestamp="2025-12-05 12:47:50 +0000 UTC" firstStartedPulling="2025-12-05 12:47:52.10027912 +0000 UTC m=+2501.594142389" lastFinishedPulling="2025-12-05 12:47:54.47984913 +0000 UTC m=+2503.973712399" observedRunningTime="2025-12-05 12:47:55.141411081 +0000 UTC m=+2504.635274360" watchObservedRunningTime="2025-12-05 12:47:55.148072954 +0000 UTC m=+2504.641936223" Dec 05 12:47:56 crc kubenswrapper[4807]: I1205 12:47:56.227246 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:56 crc kubenswrapper[4807]: I1205 12:47:56.227546 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:56 crc kubenswrapper[4807]: I1205 12:47:56.279707 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:57 crc kubenswrapper[4807]: I1205 12:47:57.198337 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:47:58 crc kubenswrapper[4807]: I1205 12:47:58.680054 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbwx4"] Dec 05 12:47:59 crc kubenswrapper[4807]: I1205 12:47:59.164597 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hbwx4" podUID="b9b0418b-c966-4dc4-9188-6201b308eaa8" containerName="registry-server" containerID="cri-o://fd96faf251d8dfc69031c1191069fc0009dd20fc042584fca5cc80b7798e0a23" gracePeriod=2 Dec 05 12:48:00 crc kubenswrapper[4807]: I1205 12:48:00.236188 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:48:00 crc kubenswrapper[4807]: E1205 12:48:00.236603 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:48:00 crc kubenswrapper[4807]: I1205 12:48:00.618349 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:48:00 crc kubenswrapper[4807]: I1205 12:48:00.620421 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:48:00 crc kubenswrapper[4807]: I1205 12:48:00.668449 4807 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.180873 4807 generic.go:334] "Generic (PLEG): container finished" podID="b9b0418b-c966-4dc4-9188-6201b308eaa8" containerID="fd96faf251d8dfc69031c1191069fc0009dd20fc042584fca5cc80b7798e0a23" exitCode=0 Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.180935 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbwx4" event={"ID":"b9b0418b-c966-4dc4-9188-6201b308eaa8","Type":"ContainerDied","Data":"fd96faf251d8dfc69031c1191069fc0009dd20fc042584fca5cc80b7798e0a23"} Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.263673 4807 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.466252 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.600265 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q4hz\" (UniqueName: \"kubernetes.io/projected/b9b0418b-c966-4dc4-9188-6201b308eaa8-kube-api-access-5q4hz\") pod \"b9b0418b-c966-4dc4-9188-6201b308eaa8\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.600412 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-utilities\") pod \"b9b0418b-c966-4dc4-9188-6201b308eaa8\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.600505 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-catalog-content\") pod \"b9b0418b-c966-4dc4-9188-6201b308eaa8\" (UID: \"b9b0418b-c966-4dc4-9188-6201b308eaa8\") " Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.605326 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-utilities" (OuterVolumeSpecName: "utilities") pod "b9b0418b-c966-4dc4-9188-6201b308eaa8" (UID: "b9b0418b-c966-4dc4-9188-6201b308eaa8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.618188 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b0418b-c966-4dc4-9188-6201b308eaa8-kube-api-access-5q4hz" (OuterVolumeSpecName: "kube-api-access-5q4hz") pod "b9b0418b-c966-4dc4-9188-6201b308eaa8" (UID: "b9b0418b-c966-4dc4-9188-6201b308eaa8"). InnerVolumeSpecName "kube-api-access-5q4hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.702951 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q4hz\" (UniqueName: \"kubernetes.io/projected/b9b0418b-c966-4dc4-9188-6201b308eaa8-kube-api-access-5q4hz\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.702987 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.715142 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9b0418b-c966-4dc4-9188-6201b308eaa8" (UID: "b9b0418b-c966-4dc4-9188-6201b308eaa8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:48:01 crc kubenswrapper[4807]: I1205 12:48:01.804284 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b0418b-c966-4dc4-9188-6201b308eaa8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:02 crc kubenswrapper[4807]: I1205 12:48:02.193489 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbwx4" Dec 05 12:48:02 crc kubenswrapper[4807]: I1205 12:48:02.193519 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbwx4" event={"ID":"b9b0418b-c966-4dc4-9188-6201b308eaa8","Type":"ContainerDied","Data":"e0ae3edd5d384542978226ef3f552c59e20397f6d93401e137cf2605e8e6c583"} Dec 05 12:48:02 crc kubenswrapper[4807]: I1205 12:48:02.193728 4807 scope.go:117] "RemoveContainer" containerID="fd96faf251d8dfc69031c1191069fc0009dd20fc042584fca5cc80b7798e0a23" Dec 05 12:48:02 crc kubenswrapper[4807]: I1205 12:48:02.230487 4807 scope.go:117] "RemoveContainer" containerID="b5f5d796d658100e98481e1c2f487ebab1dad245873b44db488c78b7ce33481e" Dec 05 12:48:02 crc kubenswrapper[4807]: I1205 12:48:02.232294 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbwx4"] Dec 05 12:48:02 crc kubenswrapper[4807]: I1205 12:48:02.241732 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hbwx4"] Dec 05 12:48:02 crc kubenswrapper[4807]: I1205 12:48:02.255885 4807 scope.go:117] "RemoveContainer" containerID="dee7f8877fb475a9e0383ea936d6907b5602ed8707beb04ef07793a9044a753b" Dec 05 12:48:03 crc kubenswrapper[4807]: I1205 12:48:03.088872 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qp6fn"] Dec 05 12:48:03 crc kubenswrapper[4807]: I1205 12:48:03.250317 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b0418b-c966-4dc4-9188-6201b308eaa8" path="/var/lib/kubelet/pods/b9b0418b-c966-4dc4-9188-6201b308eaa8/volumes" Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.209697 4807 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qp6fn" podUID="a0a51e52-ddc5-43d2-afc0-4cd69c24477a" containerName="registry-server" containerID="cri-o://2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58" gracePeriod=2 Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.664392 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.762353 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-catalog-content\") pod \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.762444 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-utilities\") pod \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.762473 4807 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h87q\" (UniqueName: \"kubernetes.io/projected/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-kube-api-access-9h87q\") pod \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\" (UID: \"a0a51e52-ddc5-43d2-afc0-4cd69c24477a\") " Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.764887 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-utilities" (OuterVolumeSpecName: "utilities") pod "a0a51e52-ddc5-43d2-afc0-4cd69c24477a" (UID: "a0a51e52-ddc5-43d2-afc0-4cd69c24477a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.768455 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-kube-api-access-9h87q" (OuterVolumeSpecName: "kube-api-access-9h87q") pod "a0a51e52-ddc5-43d2-afc0-4cd69c24477a" (UID: "a0a51e52-ddc5-43d2-afc0-4cd69c24477a"). InnerVolumeSpecName "kube-api-access-9h87q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.786653 4807 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0a51e52-ddc5-43d2-afc0-4cd69c24477a" (UID: "a0a51e52-ddc5-43d2-afc0-4cd69c24477a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.866154 4807 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.866210 4807 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:04 crc kubenswrapper[4807]: I1205 12:48:04.866220 4807 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h87q\" (UniqueName: \"kubernetes.io/projected/a0a51e52-ddc5-43d2-afc0-4cd69c24477a-kube-api-access-9h87q\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.224358 4807 generic.go:334] "Generic (PLEG): container finished" podID="a0a51e52-ddc5-43d2-afc0-4cd69c24477a" containerID="2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58" exitCode=0 Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.224723 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qp6fn" event={"ID":"a0a51e52-ddc5-43d2-afc0-4cd69c24477a","Type":"ContainerDied","Data":"2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58"} Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.224768 4807 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qp6fn" event={"ID":"a0a51e52-ddc5-43d2-afc0-4cd69c24477a","Type":"ContainerDied","Data":"8e636f497c3ea13c491258f82924ae7e8d9b33e8fc8fe90e65e39fb122b8a8ff"} Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.224789 4807 scope.go:117] "RemoveContainer" containerID="2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.224947 4807 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qp6fn" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.266622 4807 scope.go:117] "RemoveContainer" containerID="2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.278313 4807 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qp6fn"] Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.284977 4807 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qp6fn"] Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.297739 4807 scope.go:117] "RemoveContainer" containerID="41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.336650 4807 scope.go:117] "RemoveContainer" containerID="2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58" Dec 05 12:48:05 crc kubenswrapper[4807]: E1205 12:48:05.338637 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58\": container with ID starting with 2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58 not found: ID does not exist" containerID="2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.338678 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58"} err="failed to get container status \"2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58\": rpc error: code = NotFound desc = could not find container \"2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58\": container with ID starting with 2f9d58bbb1fcba2950374cf65e01862376399e57f134010ca01f1e75e8cd9d58 not found: ID does not exist" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.338728 4807 scope.go:117] "RemoveContainer" containerID="2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7" Dec 05 12:48:05 crc kubenswrapper[4807]: E1205 12:48:05.339282 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7\": container with ID starting with 2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7 not found: ID does not exist" containerID="2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.339324 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7"} err="failed to get container status \"2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7\": rpc error: code = NotFound desc = could not find container \"2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7\": container with ID starting with 2e99680d0ded0fc28e1f22159bc01dee4a769ad9e1c2768a7dc255073e5fbeb7 not found: ID does not exist" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.339345 4807 scope.go:117] "RemoveContainer" containerID="41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b" Dec 05 12:48:05 crc kubenswrapper[4807]: E1205 12:48:05.339588 4807 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b\": container with ID starting with 41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b not found: ID does not exist" containerID="41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b" Dec 05 12:48:05 crc kubenswrapper[4807]: I1205 12:48:05.339632 4807 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b"} err="failed to get container status \"41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b\": rpc error: code = NotFound desc = could not find container \"41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b\": container with ID starting with 41b58a891f9967eee76be4a16282fa2a9a29ad5465f8d28afb35d12b0c480f4b not found: ID does not exist" Dec 05 12:48:07 crc kubenswrapper[4807]: I1205 12:48:07.246988 4807 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0a51e52-ddc5-43d2-afc0-4cd69c24477a" path="/var/lib/kubelet/pods/a0a51e52-ddc5-43d2-afc0-4cd69c24477a/volumes" Dec 05 12:48:12 crc kubenswrapper[4807]: I1205 12:48:12.235902 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:48:12 crc kubenswrapper[4807]: E1205 12:48:12.236392 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:48:23 crc kubenswrapper[4807]: I1205 12:48:23.236035 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:48:23 crc kubenswrapper[4807]: E1205 12:48:23.236788 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" Dec 05 12:48:35 crc kubenswrapper[4807]: I1205 12:48:35.235937 4807 scope.go:117] "RemoveContainer" containerID="ba51fd4057357c6a8758cf3049ac0ab5bd78ae4470f8ec01dd759805332f1d9c" Dec 05 12:48:35 crc kubenswrapper[4807]: E1205 12:48:35.236848 4807 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kth9r_openshift-machine-config-operator(a779882f-2b7e-4ae0-addd-686fd4343bb6)\"" pod="openshift-machine-config-operator/machine-config-daemon-kth9r" podUID="a779882f-2b7e-4ae0-addd-686fd4343bb6" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114552260024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114552261017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114544715016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114544715015464 5ustar corecore